var/home/core/zuul-output/0000755000175000017500000000000015114222460014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114234212015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004643720015114234203017676 0ustar rootrootDec 04 06:11:01 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 06:11:01 crc restorecon[4684]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:01 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 06:11:02 crc restorecon[4684]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 06:11:02 crc kubenswrapper[4685]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 06:11:02 crc kubenswrapper[4685]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 06:11:02 crc kubenswrapper[4685]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 06:11:02 crc kubenswrapper[4685]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 06:11:02 crc kubenswrapper[4685]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 06:11:02 crc kubenswrapper[4685]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.944225 4685 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947270 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947288 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947293 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947298 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947302 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947306 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947310 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947314 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947319 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947322 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947327 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947337 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947342 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947346 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947350 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947355 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947360 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947365 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947370 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947374 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947379 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947384 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947389 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947394 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947398 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947402 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947419 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947423 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947427 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947430 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947434 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947437 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947441 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947445 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947449 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947453 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947457 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947461 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947465 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947468 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947472 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947475 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947479 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947483 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947486 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947490 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947495 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947500 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947505 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947509 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947513 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947517 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947521 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947525 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947529 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947532 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947536 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947540 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947544 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947548 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947553 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947556 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947560 4685 feature_gate.go:330] unrecognized feature gate: Example Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947564 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947568 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947572 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947575 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947579 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947583 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947587 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.947590 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948117 4685 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948127 4685 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948135 4685 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948141 4685 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948146 4685 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948150 4685 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948156 4685 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948162 4685 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948167 4685 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948171 4685 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948175 4685 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948179 4685 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948184 4685 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948188 4685 flags.go:64] FLAG: --cgroup-root="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948195 4685 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948199 4685 flags.go:64] FLAG: --client-ca-file="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948203 4685 flags.go:64] FLAG: --cloud-config="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948207 4685 flags.go:64] FLAG: --cloud-provider="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948211 4685 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948216 4685 flags.go:64] FLAG: --cluster-domain="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948220 4685 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948225 4685 flags.go:64] FLAG: --config-dir="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948229 4685 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948234 4685 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948240 4685 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948244 4685 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948248 4685 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948252 4685 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948256 4685 flags.go:64] FLAG: --contention-profiling="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948261 4685 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948265 4685 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948270 4685 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948274 4685 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948279 4685 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948283 4685 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948287 4685 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948291 4685 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948295 4685 flags.go:64] FLAG: --enable-server="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948299 4685 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948309 4685 flags.go:64] FLAG: --event-burst="100" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948313 4685 flags.go:64] FLAG: --event-qps="50" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948318 4685 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948322 4685 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948326 4685 flags.go:64] FLAG: --eviction-hard="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948331 4685 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948336 4685 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948341 4685 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948346 4685 flags.go:64] FLAG: --eviction-soft="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948350 4685 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948354 4685 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948359 4685 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948363 4685 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948366 4685 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948370 4685 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948374 4685 flags.go:64] FLAG: --feature-gates="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948380 4685 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948384 4685 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948388 4685 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948392 4685 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948396 4685 flags.go:64] FLAG: --healthz-port="10248" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948400 4685 flags.go:64] FLAG: --help="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948429 4685 flags.go:64] FLAG: --hostname-override="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948434 4685 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948438 4685 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948443 4685 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948447 4685 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948452 4685 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948457 4685 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948462 4685 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948466 4685 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948470 4685 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948475 4685 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948480 4685 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948484 4685 flags.go:64] FLAG: --kube-reserved="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948488 4685 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948492 4685 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948496 4685 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948500 4685 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948506 4685 flags.go:64] FLAG: --lock-file="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948509 4685 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948514 4685 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948517 4685 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948524 4685 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948528 4685 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948532 4685 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948536 4685 flags.go:64] FLAG: --logging-format="text" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948540 4685 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948544 4685 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948548 4685 flags.go:64] FLAG: --manifest-url="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948552 4685 flags.go:64] FLAG: --manifest-url-header="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948558 4685 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948562 4685 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948567 4685 flags.go:64] FLAG: --max-pods="110" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948571 4685 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948575 4685 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948579 4685 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948583 4685 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948587 4685 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948591 4685 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948595 4685 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948605 4685 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948609 4685 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948613 4685 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948617 4685 flags.go:64] FLAG: --pod-cidr="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948621 4685 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948628 4685 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948631 4685 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948636 4685 flags.go:64] FLAG: --pods-per-core="0" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948640 4685 flags.go:64] FLAG: --port="10250" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948644 4685 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948648 4685 flags.go:64] FLAG: --provider-id="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948652 4685 flags.go:64] FLAG: --qos-reserved="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948657 4685 flags.go:64] FLAG: --read-only-port="10255" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948661 4685 flags.go:64] FLAG: --register-node="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948665 4685 flags.go:64] FLAG: --register-schedulable="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948669 4685 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948676 4685 flags.go:64] FLAG: --registry-burst="10" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948680 4685 flags.go:64] FLAG: --registry-qps="5" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948684 4685 flags.go:64] FLAG: --reserved-cpus="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948688 4685 flags.go:64] FLAG: --reserved-memory="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948693 4685 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948698 4685 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948702 4685 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948706 4685 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948709 4685 flags.go:64] FLAG: --runonce="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948713 4685 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948717 4685 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948721 4685 flags.go:64] FLAG: --seccomp-default="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948726 4685 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948730 4685 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948734 4685 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948738 4685 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948742 4685 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948747 4685 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948751 4685 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948755 4685 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948758 4685 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948763 4685 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948767 4685 flags.go:64] FLAG: --system-cgroups="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948771 4685 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948777 4685 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948781 4685 flags.go:64] FLAG: --tls-cert-file="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948786 4685 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948791 4685 flags.go:64] FLAG: --tls-min-version="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948795 4685 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948799 4685 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948803 4685 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948807 4685 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948812 4685 flags.go:64] FLAG: --v="2" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948820 4685 flags.go:64] FLAG: --version="false" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948826 4685 flags.go:64] FLAG: --vmodule="" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948831 4685 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.948836 4685 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948967 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948972 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948977 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948982 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948987 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948991 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948995 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.948998 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949002 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949005 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949010 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949014 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949018 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949022 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949026 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949030 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949034 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949037 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949041 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949045 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949048 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949052 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949056 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949060 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949063 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949067 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949070 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949074 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949077 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949081 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949085 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949089 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949093 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949097 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949100 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949103 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949107 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949110 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949114 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949121 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949125 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949128 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949132 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949135 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949139 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949142 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949145 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949149 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949152 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949156 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949159 4685 feature_gate.go:330] unrecognized feature gate: Example Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949162 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949166 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949169 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949173 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949177 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949180 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949185 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949189 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949193 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949196 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949200 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949204 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949207 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949211 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949214 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949218 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949221 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949225 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949229 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.949232 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.949438 4685 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.962859 4685 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.962883 4685 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962946 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962952 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962957 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962962 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962966 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962970 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962974 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962978 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962982 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962985 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962989 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.962994 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963000 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963006 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963011 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963015 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963021 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963027 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963031 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963035 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963039 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963043 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963047 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963052 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963056 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963059 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963063 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963067 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963070 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963074 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963079 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963083 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963086 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963090 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963093 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963097 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963101 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963105 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963108 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963112 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963116 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963121 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963125 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963129 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963133 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963137 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963141 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963145 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963148 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963151 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963155 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963159 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963163 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963166 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963171 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963174 4685 feature_gate.go:330] unrecognized feature gate: Example Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963178 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963181 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963185 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963189 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963193 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963196 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963200 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963203 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963207 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963210 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963215 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963219 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963223 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963228 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963231 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.963238 4685 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963344 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963351 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963354 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963358 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963361 4685 feature_gate.go:330] unrecognized feature gate: Example Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963365 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963368 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963373 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963377 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963380 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963384 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963387 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963391 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963394 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963399 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963404 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963426 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963432 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963438 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963442 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963446 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963449 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963453 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963456 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963460 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963464 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963467 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963471 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963474 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963477 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963481 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963484 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963488 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963492 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963495 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963499 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963502 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963506 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963510 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963514 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963518 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963522 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963526 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963529 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963533 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963536 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963540 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963543 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963547 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963551 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963556 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963560 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963564 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963568 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963572 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963576 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963580 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963585 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963589 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963593 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963597 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963601 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963605 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963610 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963613 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963618 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963623 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963628 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963632 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963636 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 06:11:02 crc kubenswrapper[4685]: W1204 06:11:02.963640 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.963647 4685 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.963791 4685 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.966585 4685 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.966673 4685 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.967179 4685 server.go:997] "Starting client certificate rotation" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.967194 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.967732 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-03 22:19:16.498156893 +0000 UTC Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.967815 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.972671 4685 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 06:11:02 crc kubenswrapper[4685]: E1204 06:11:02.974016 4685 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.974188 4685 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 06:11:02 crc kubenswrapper[4685]: I1204 06:11:02.982541 4685 log.go:25] "Validated CRI v1 runtime API" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.004278 4685 log.go:25] "Validated CRI v1 image API" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.007063 4685 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.009582 4685 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-06-06-19-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.009613 4685 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.030547 4685 manager.go:217] Machine: {Timestamp:2025-12-04 06:11:03.028035623 +0000 UTC m=+0.216266478 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:83da537d-e4a5-4b93-8404-e36573fe734d BootID:19f8f547-1e23-412e-9678-012276047599 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:d3:38:49 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:d3:38:49 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f6:10:51 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c2:3a:9e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e5:db:d3 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:cd:14:88 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:78:78:37:96:33 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ca:6f:2a:87:3b:9c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.031076 4685 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.031324 4685 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.032039 4685 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.032390 4685 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.032495 4685 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.032844 4685 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.032864 4685 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.033217 4685 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.033273 4685 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.033647 4685 state_mem.go:36] "Initialized new in-memory state store" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.033930 4685 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.034875 4685 kubelet.go:418] "Attempting to sync node with API server" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.034921 4685 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.034963 4685 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.034984 4685 kubelet.go:324] "Adding apiserver pod source" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.035066 4685 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.036580 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.036619 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.036699 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.036705 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.037227 4685 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.037801 4685 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.038898 4685 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039712 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039755 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039773 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039788 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039812 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039826 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039841 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039866 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039882 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039896 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039914 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.039930 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.040243 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.040992 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.041669 4685 server.go:1280] "Started kubelet" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.042254 4685 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.042163 4685 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 06:11:03 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045003 4685 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045093 4685 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.045258 4685 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045227 4685 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 14:04:52.791768367 +0000 UTC Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045289 4685 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045359 4685 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045301 4685 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 967h53m49.746471148s for next certificate rotation Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.045714 4685 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.046654 4685 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.046811 4685 factory.go:55] Registering systemd factory Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.046844 4685 factory.go:221] Registration of the systemd container factory successfully Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.047482 4685 factory.go:153] Registering CRI-O factory Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.047558 4685 factory.go:221] Registration of the crio container factory successfully Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.046091 4685 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dee518e05409f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 06:11:03.041573023 +0000 UTC m=+0.229803798,LastTimestamp:2025-12-04 06:11:03.041573023 +0000 UTC m=+0.229803798,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.047498 4685 server.go:460] "Adding debug handlers to kubelet server" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.047776 4685 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.047480 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="200ms" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.047890 4685 factory.go:103] Registering Raw factory Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.047928 4685 manager.go:1196] Started watching for new ooms in manager Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.049602 4685 manager.go:319] Starting recovery of all containers Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.053156 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.053319 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067514 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067587 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067606 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067621 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067640 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067652 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067665 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067679 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067694 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067713 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067728 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067744 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067760 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067779 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067792 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067805 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067845 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067859 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067878 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067894 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067918 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067933 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067946 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067959 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067971 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.067989 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068011 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068025 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068038 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068051 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068063 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068080 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068097 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068110 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068122 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068135 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068146 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068159 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068172 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068184 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068196 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068209 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068222 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068235 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068248 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068261 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068274 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068287 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068371 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068390 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068447 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068461 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068478 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068499 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068512 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068530 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068547 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068562 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068576 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068588 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068606 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068619 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068634 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068647 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068661 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068695 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068710 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068724 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068743 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068756 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068769 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068781 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068795 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068810 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068830 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068868 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068884 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068901 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068926 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068941 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068954 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068967 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068979 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.068991 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069003 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069015 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069027 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069041 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069052 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069065 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069085 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069104 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069117 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069129 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069142 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069155 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069168 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069182 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069196 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069208 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069223 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069236 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069249 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069268 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069288 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069302 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069316 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069335 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069348 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069363 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069377 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069390 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069421 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069437 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069449 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069462 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069475 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069510 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069528 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069568 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069585 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069597 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069612 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069626 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069639 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069653 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069665 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069677 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069691 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069708 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069721 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069752 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069773 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069787 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069800 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069812 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069825 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069837 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069850 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069861 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069874 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069891 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069903 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069916 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069935 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069947 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069960 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069975 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.069989 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070004 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070016 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070036 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070053 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070069 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070087 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070104 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070119 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070134 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070151 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070165 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070177 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070190 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070202 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070215 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070230 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070243 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070263 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070277 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070291 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.070304 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073246 4685 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073288 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073305 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073321 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073336 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073349 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073372 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073386 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073399 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073443 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073462 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073475 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073488 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073501 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073519 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073531 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073548 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073560 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073572 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073589 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073607 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073665 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073677 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073691 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073719 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073762 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073802 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073832 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073844 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073856 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073869 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073887 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073903 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073916 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073961 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073976 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.073992 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.074011 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.074023 4685 reconstruct.go:97] "Volume reconstruction finished" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.074032 4685 reconciler.go:26] "Reconciler: start to sync state" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.082158 4685 manager.go:324] Recovery completed Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.093298 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.095305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.095613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.095624 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.096597 4685 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.096616 4685 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.096637 4685 state_mem.go:36] "Initialized new in-memory state store" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.104994 4685 policy_none.go:49] "None policy: Start" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.107593 4685 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.107677 4685 state_mem.go:35] "Initializing new in-memory state store" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.120635 4685 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.124242 4685 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.124314 4685 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.124422 4685 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.124497 4685 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.125436 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.125507 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.147648 4685 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.159949 4685 manager.go:334] "Starting Device Plugin manager" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.160093 4685 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.160124 4685 server.go:79] "Starting device plugin registration server" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.160888 4685 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.160916 4685 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.161306 4685 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.161436 4685 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.161447 4685 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.179052 4685 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.225531 4685 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.225721 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.227338 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.227474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.227502 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.227786 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.228072 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.228124 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229246 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229286 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229571 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229759 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.229811 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.230780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.230812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.230845 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.230844 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.230896 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.230867 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231084 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231154 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231184 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231813 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231815 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231824 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231839 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.231897 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.232146 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.232213 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.233039 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.233058 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.233072 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.233287 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.233314 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.234815 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.234841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.234851 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.234891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.234908 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.234918 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.248396 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="400ms" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.261447 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.262232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.262260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.262269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.262304 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.262680 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276212 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276276 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276340 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276366 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276449 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276504 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276535 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276584 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276625 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276660 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276684 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276726 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276763 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276795 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.276815 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378012 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378099 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378138 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378177 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378215 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378250 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378257 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378304 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378284 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378371 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378515 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378571 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378574 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378673 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378717 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378750 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378781 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378796 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378810 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378848 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378862 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378883 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378747 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378928 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378932 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.378982 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.379002 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.379049 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.462907 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.465630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.465708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.465731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.465774 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.466535 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.560522 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.568050 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.590847 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e8987b750fe01781ad8bf256c77d0ff509831ad9cb9a4be5ebce9b470cf877d8 WatchSource:0}: Error finding container e8987b750fe01781ad8bf256c77d0ff509831ad9cb9a4be5ebce9b470cf877d8: Status 404 returned error can't find the container with id e8987b750fe01781ad8bf256c77d0ff509831ad9cb9a4be5ebce9b470cf877d8 Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.592174 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.593962 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4f866b4f38594d655fd350e1ac5de76fc51f92ab8143138f9bb2b5b9f94e55b6 WatchSource:0}: Error finding container 4f866b4f38594d655fd350e1ac5de76fc51f92ab8143138f9bb2b5b9f94e55b6: Status 404 returned error can't find the container with id 4f866b4f38594d655fd350e1ac5de76fc51f92ab8143138f9bb2b5b9f94e55b6 Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.606599 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-8f1253783eba04d69a33c306687b7ad79c92898910254e7ea8e533f03e91f80a WatchSource:0}: Error finding container 8f1253783eba04d69a33c306687b7ad79c92898910254e7ea8e533f03e91f80a: Status 404 returned error can't find the container with id 8f1253783eba04d69a33c306687b7ad79c92898910254e7ea8e533f03e91f80a Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.620345 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.625741 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.643183 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-92fe691d57a48c6ef79c68ec034d1f85f89f54bd71cc97d2f69bb23f8b5fafe6 WatchSource:0}: Error finding container 92fe691d57a48c6ef79c68ec034d1f85f89f54bd71cc97d2f69bb23f8b5fafe6: Status 404 returned error can't find the container with id 92fe691d57a48c6ef79c68ec034d1f85f89f54bd71cc97d2f69bb23f8b5fafe6 Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.648378 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e83635532f11d6ec68d8d49a3dbd2e1d2a0102ec3752efc7af5ec7ae791b4bdd WatchSource:0}: Error finding container e83635532f11d6ec68d8d49a3dbd2e1d2a0102ec3752efc7af5ec7ae791b4bdd: Status 404 returned error can't find the container with id e83635532f11d6ec68d8d49a3dbd2e1d2a0102ec3752efc7af5ec7ae791b4bdd Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.650531 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="800ms" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.866832 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.868048 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.868084 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.868096 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:03 crc kubenswrapper[4685]: I1204 06:11:03.868119 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.868508 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.943874 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.943942 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:03 crc kubenswrapper[4685]: W1204 06:11:03.984626 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:03 crc kubenswrapper[4685]: E1204 06:11:03.985850 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.042311 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.128576 4685 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa" exitCode=0 Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.128681 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.128762 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e83635532f11d6ec68d8d49a3dbd2e1d2a0102ec3752efc7af5ec7ae791b4bdd"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.128839 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130063 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130149 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130382 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624" exitCode=0 Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130449 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130467 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"92fe691d57a48c6ef79c68ec034d1f85f89f54bd71cc97d2f69bb23f8b5fafe6"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.130524 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131716 4685 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b488be16444eb1eb7491e50f313a22c589e3c97b5a32ecb4214d897116cb88f8" exitCode=0 Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131761 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b488be16444eb1eb7491e50f313a22c589e3c97b5a32ecb4214d897116cb88f8"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131776 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8f1253783eba04d69a33c306687b7ad79c92898910254e7ea8e533f03e91f80a"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.131837 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.132421 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.132449 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.132460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.134096 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.134848 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.134888 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4f866b4f38594d655fd350e1ac5de76fc51f92ab8143138f9bb2b5b9f94e55b6"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.135316 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.135383 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.135397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.136087 4685 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="85ea114148146cd230facc520dc234be71720f9113ffeb63105773253b994b83" exitCode=0 Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.136116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"85ea114148146cd230facc520dc234be71720f9113ffeb63105773253b994b83"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.136131 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e8987b750fe01781ad8bf256c77d0ff509831ad9cb9a4be5ebce9b470cf877d8"} Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.136178 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.137187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.137220 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.137232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:04 crc kubenswrapper[4685]: E1204 06:11:04.452483 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="1.6s" Dec 04 06:11:04 crc kubenswrapper[4685]: W1204 06:11:04.546326 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:04 crc kubenswrapper[4685]: E1204 06:11:04.546401 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.669186 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.672876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.672901 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.672911 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:04 crc kubenswrapper[4685]: I1204 06:11:04.672933 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 06:11:04 crc kubenswrapper[4685]: E1204 06:11:04.673280 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Dec 04 06:11:04 crc kubenswrapper[4685]: W1204 06:11:04.673335 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Dec 04 06:11:04 crc kubenswrapper[4685]: E1204 06:11:04.673398 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.142350 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.142425 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.142441 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.142456 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.142469 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.142600 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.143613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.143646 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.143656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.144227 4685 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e39343d51bdfc3bfea0729b6c3053a1fda3d152c9598821175a9bc7224d4ea06" exitCode=0 Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.144297 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e39343d51bdfc3bfea0729b6c3053a1fda3d152c9598821175a9bc7224d4ea06"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.144455 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.145361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.145425 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.145439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.148809 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.148847 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.148861 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.148904 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.149967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.150000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.150012 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.152426 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"03665be2d8de803b89268c93e59daa31ebba2c1cefc03f9c7864341c011aee0e"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.152523 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.153247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.153283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.153295 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.155783 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.155822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.155836 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81"} Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.156024 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.157155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.157187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.157200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.166725 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 06:11:05 crc kubenswrapper[4685]: I1204 06:11:05.533595 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.034254 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.163463 4685 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="749f108b5d2c0cb1b59004cd9eed7b086ed8c114b785cd7b8ac3b39360bfeccc" exitCode=0 Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.163611 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"749f108b5d2c0cb1b59004cd9eed7b086ed8c114b785cd7b8ac3b39360bfeccc"} Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.163689 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.163707 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.163763 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.163814 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.164546 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.164603 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165762 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165833 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165923 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.165945 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.166585 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.166631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.166652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.273899 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.275653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.275714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.275732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:06 crc kubenswrapper[4685]: I1204 06:11:06.275768 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.169070 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f53c3fb82ec9be442573ff7dcfe1c79afa0f953247cde760f548c14c6e04a00e"} Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.169119 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.169259 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.169346 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.169123 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4eaf576a68effea89d18dfd61e57925eaf620301092b9e754baed66bea6b2c29"} Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.169526 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8b611f3d4f9990a787ec8ad0d0a986bb48ca34634767233da48dc9bce5c226ba"} Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.170523 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.170576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.170598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.170891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.170931 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:07 crc kubenswrapper[4685]: I1204 06:11:07.170947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.180256 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1908d1bb91749e479fd500f882eb0dbbe62e2e713b7f9276eff32f90a3b16076"} Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.180338 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"abbb530cb50d955a00b7b028721937f1e522d6285eb821f88628cd507cf7eff2"} Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.180489 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.181786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.181847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.181872 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.185512 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.915810 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.915954 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.915993 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.917099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.917130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:08 crc kubenswrapper[4685]: I1204 06:11:08.917140 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:09 crc kubenswrapper[4685]: I1204 06:11:09.183919 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:09 crc kubenswrapper[4685]: I1204 06:11:09.186152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:09 crc kubenswrapper[4685]: I1204 06:11:09.186229 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:09 crc kubenswrapper[4685]: I1204 06:11:09.186250 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:09 crc kubenswrapper[4685]: I1204 06:11:09.631183 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 06:11:10 crc kubenswrapper[4685]: I1204 06:11:10.187058 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:10 crc kubenswrapper[4685]: I1204 06:11:10.188481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:10 crc kubenswrapper[4685]: I1204 06:11:10.188560 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:10 crc kubenswrapper[4685]: I1204 06:11:10.188578 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:11 crc kubenswrapper[4685]: I1204 06:11:11.189287 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:11 crc kubenswrapper[4685]: I1204 06:11:11.190391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:11 crc kubenswrapper[4685]: I1204 06:11:11.190461 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:11 crc kubenswrapper[4685]: I1204 06:11:11.190476 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.254110 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.254345 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.255799 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.255863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.255880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.768181 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.768485 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.769978 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.770031 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:12 crc kubenswrapper[4685]: I1204 06:11:12.770050 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:13 crc kubenswrapper[4685]: E1204 06:11:13.179935 4685 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.491254 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.491601 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.493262 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.493315 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.493335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.574004 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.574216 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.575741 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.575782 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:13 crc kubenswrapper[4685]: I1204 06:11:13.575794 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.111862 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.119704 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.196280 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.197270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.197307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.197318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:14 crc kubenswrapper[4685]: I1204 06:11:14.201002 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.042965 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 06:11:15 crc kubenswrapper[4685]: E1204 06:11:15.168441 4685 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.199127 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.200245 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.200276 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.200289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.534003 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.534122 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.768796 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 06:11:15 crc kubenswrapper[4685]: I1204 06:11:15.768903 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 06:11:16 crc kubenswrapper[4685]: E1204 06:11:16.053543 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.202386 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.203806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.203913 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.203936 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:16 crc kubenswrapper[4685]: W1204 06:11:16.240483 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.240598 4685 trace.go:236] Trace[1610993214]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 06:11:06.238) (total time: 10002ms): Dec 04 06:11:16 crc kubenswrapper[4685]: Trace[1610993214]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:11:16.240) Dec 04 06:11:16 crc kubenswrapper[4685]: Trace[1610993214]: [10.002091244s] [10.002091244s] END Dec 04 06:11:16 crc kubenswrapper[4685]: E1204 06:11:16.240632 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 06:11:16 crc kubenswrapper[4685]: W1204 06:11:16.256651 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.256736 4685 trace.go:236] Trace[1901742203]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 06:11:06.254) (total time: 10001ms): Dec 04 06:11:16 crc kubenswrapper[4685]: Trace[1901742203]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:11:16.256) Dec 04 06:11:16 crc kubenswrapper[4685]: Trace[1901742203]: [10.001849294s] [10.001849294s] END Dec 04 06:11:16 crc kubenswrapper[4685]: E1204 06:11:16.256758 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 06:11:16 crc kubenswrapper[4685]: E1204 06:11:16.277348 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 04 06:11:16 crc kubenswrapper[4685]: W1204 06:11:16.455120 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 06:11:16 crc kubenswrapper[4685]: I1204 06:11:16.455271 4685 trace.go:236] Trace[946455752]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 06:11:06.453) (total time: 10001ms): Dec 04 06:11:16 crc kubenswrapper[4685]: Trace[946455752]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:11:16.455) Dec 04 06:11:16 crc kubenswrapper[4685]: Trace[946455752]: [10.001945301s] [10.001945301s] END Dec 04 06:11:16 crc kubenswrapper[4685]: E1204 06:11:16.455312 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 06:11:17 crc kubenswrapper[4685]: W1204 06:11:17.699922 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 06:11:17 crc kubenswrapper[4685]: I1204 06:11:17.700103 4685 trace.go:236] Trace[983247868]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 06:11:07.698) (total time: 10001ms): Dec 04 06:11:17 crc kubenswrapper[4685]: Trace[983247868]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:11:17.699) Dec 04 06:11:17 crc kubenswrapper[4685]: Trace[983247868]: [10.001952027s] [10.001952027s] END Dec 04 06:11:17 crc kubenswrapper[4685]: E1204 06:11:17.700145 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 06:11:18 crc kubenswrapper[4685]: E1204 06:11:18.220320 4685 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187dee518e05409f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 06:11:03.041573023 +0000 UTC m=+0.229803798,LastTimestamp:2025-12-04 06:11:03.041573023 +0000 UTC m=+0.229803798,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 06:11:18 crc kubenswrapper[4685]: I1204 06:11:18.221676 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 06:11:18 crc kubenswrapper[4685]: I1204 06:11:18.221906 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:18 crc kubenswrapper[4685]: I1204 06:11:18.223704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:18 crc kubenswrapper[4685]: I1204 06:11:18.223750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:18 crc kubenswrapper[4685]: I1204 06:11:18.223767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:18 crc kubenswrapper[4685]: I1204 06:11:18.237581 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.210157 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.211493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.211557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.211570 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.267467 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.477551 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.479725 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.479779 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.479797 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:19 crc kubenswrapper[4685]: I1204 06:11:19.479833 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 06:11:22 crc kubenswrapper[4685]: I1204 06:11:22.134604 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 06:11:22 crc kubenswrapper[4685]: I1204 06:11:22.134689 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 06:11:22 crc kubenswrapper[4685]: I1204 06:11:22.140786 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 06:11:22 crc kubenswrapper[4685]: I1204 06:11:22.140898 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 06:11:23 crc kubenswrapper[4685]: E1204 06:11:23.180986 4685 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.543338 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.543663 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.545207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.545290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.545315 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.549147 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.769070 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 06:11:25 crc kubenswrapper[4685]: I1204 06:11:25.769156 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 06:11:26 crc kubenswrapper[4685]: I1204 06:11:26.226637 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:11:26 crc kubenswrapper[4685]: I1204 06:11:26.226696 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 06:11:26 crc kubenswrapper[4685]: I1204 06:11:26.227890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:26 crc kubenswrapper[4685]: I1204 06:11:26.227920 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:26 crc kubenswrapper[4685]: I1204 06:11:26.227931 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.151877 4685 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.152359 4685 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.152961 4685 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.155656 4685 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.158991 4685 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.161553 4685 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.161820 4685 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.163032 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.163063 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.163071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.163086 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.163094 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.166982 4685 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 04 06:11:27 crc kubenswrapper[4685]: E1204 06:11:27.187873 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.194955 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57136->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.195045 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57136->192.168.126.11:17697: read: connection reset by peer" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.194957 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57150->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.195143 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57150->192.168.126.11:17697: read: connection reset by peer" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.195928 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.196041 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.196083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.196126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.196146 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.196175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.196199 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: E1204 06:11:27.210721 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.214929 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.214966 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.214976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.214991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.215001 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: E1204 06:11:27.253860 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.259621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.259653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.259664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.259707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.259721 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: E1204 06:11:27.273765 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.278281 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.278309 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.278317 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.278329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.278337 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: E1204 06:11:27.289942 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:27 crc kubenswrapper[4685]: E1204 06:11:27.290055 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.291458 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.291481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.291489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.291502 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.291511 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.394220 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.394257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.394266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.394283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.394292 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.497356 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.497458 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.497471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.497499 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.497515 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.600593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.600650 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.600664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.600684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.600698 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.704309 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.704362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.704380 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.704436 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.704454 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.807510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.807581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.807600 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.807624 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.807641 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.910949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.911003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.911014 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.911037 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:27 crc kubenswrapper[4685]: I1204 06:11:27.911049 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:27Z","lastTransitionTime":"2025-12-04T06:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.014345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.014483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.014505 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.014532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.014548 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.051445 4685 apiserver.go:52] "Watching apiserver" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.055584 4685 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.055906 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.056379 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.056580 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.056391 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.056950 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.057103 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.057135 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.056973 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.057381 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.057506 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.059234 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.060659 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.060808 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.060863 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.061214 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.061229 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.061363 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.061988 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.064383 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.094566 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.113627 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.118474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.118556 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.118598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.118630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.118653 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.130745 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.144980 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.147931 4685 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.162484 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.164829 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.164909 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.165006 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.165055 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.165562 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.165638 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.165715 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.165809 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.166228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.166343 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.166550 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.166565 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.166472 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167293 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167045 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167360 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167477 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167501 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167581 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167584 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167629 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167677 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167722 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167770 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167815 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167860 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167952 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168004 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.167863 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168005 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168055 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168127 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168182 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168203 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168228 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168283 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168333 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168590 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168649 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168700 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168739 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168757 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168797 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168840 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168886 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168915 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168932 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.168954 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.168981 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:11:28.668943449 +0000 UTC m=+25.857174274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169050 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169083 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169133 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169199 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169243 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169298 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169337 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169388 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169348 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169481 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169498 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169551 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169585 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169618 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169615 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169682 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169715 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169912 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.169962 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170000 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170031 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170079 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170146 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170171 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170222 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170269 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170304 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170336 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170369 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170400 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170524 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170558 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170594 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170629 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170662 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170706 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170739 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170771 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170803 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170837 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170878 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170928 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170913 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.170978 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171016 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171058 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171121 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171182 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171189 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171215 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171247 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171277 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171308 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171338 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171370 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171403 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171470 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171507 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171546 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171610 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171660 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171696 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171673 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171745 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171783 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171821 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171860 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171891 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171923 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.171963 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172002 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172056 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172104 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172149 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172192 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172237 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172286 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172330 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172377 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172325 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172494 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172521 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172609 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172650 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172684 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172714 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172746 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172784 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172816 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172849 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172882 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172916 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172950 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.172982 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173014 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173050 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173083 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173117 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173148 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173180 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173205 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173213 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173295 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173313 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173376 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173460 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173476 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173504 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173545 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173560 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173593 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174073 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174363 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174450 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174521 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174556 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174609 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.173581 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174888 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.174953 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175053 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175092 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175129 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175163 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175201 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175193 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175238 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175243 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175315 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175275 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175398 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175498 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175534 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175565 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175592 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175615 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175638 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175662 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175685 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175714 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175736 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175759 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175783 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175805 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175828 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175852 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175877 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175900 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175930 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175955 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175978 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.175971 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176008 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176156 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176171 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176213 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176260 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176327 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176386 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176478 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176524 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176638 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176679 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176729 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176780 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176837 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176891 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176951 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177007 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177052 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177086 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177128 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177188 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177255 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177303 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177351 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177396 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177520 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177558 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177595 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177629 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177664 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177700 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177737 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177772 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177807 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177841 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177878 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177944 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177985 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178041 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178098 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178198 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178242 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178276 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178319 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178354 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178390 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179690 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179736 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179772 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179808 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179903 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179928 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179951 4685 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179973 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179992 4685 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180014 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180034 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180054 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180074 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180094 4685 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180115 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180135 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180155 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180174 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180194 4685 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180213 4685 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180233 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180255 4685 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180277 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180297 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180449 4685 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180472 4685 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180494 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180514 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180592 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180633 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180662 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180690 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180719 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180747 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180775 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180810 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180832 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180853 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180873 4685 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180892 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180911 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180929 4685 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180951 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180972 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180992 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181012 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181031 4685 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181051 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181070 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181090 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.183082 4685 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.187300 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176885 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.176945 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177121 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177166 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177566 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.177599 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.178636 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179199 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179203 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179552 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.179571 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180213 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180475 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180566 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.180890 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181141 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181166 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.181704 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182049 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182075 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182425 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182517 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182667 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182670 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182688 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.182925 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.183336 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.183733 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.183845 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.183915 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184047 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184089 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184390 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184537 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184805 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184838 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.184996 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.185679 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.185848 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.185978 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.185958 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.186342 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.186472 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.186588 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.186679 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.186726 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.186971 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.187071 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.187272 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.188064 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.188168 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.189268 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.189311 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.189327 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.189496 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.189864 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.190046 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.191671 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.194573 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.194759 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.194898 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195077 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195141 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195252 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195584 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195712 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.195824 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.196303 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.196326 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.198160 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.198940 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.199535 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.200030 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.200667 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.200715 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.200795 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201145 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201506 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201525 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201626 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201736 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201800 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.201776 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.202098 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.202122 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.202600 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.202621 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.202643 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.202754 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.203217 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.203338 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.203485 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.203511 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:28.703471008 +0000 UTC m=+25.891701873 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.203669 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.203764 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.203775 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:28.703754047 +0000 UTC m=+25.891984972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.203931 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.204183 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.204740 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.204980 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.205392 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.207798 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.208131 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.208463 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.208604 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.209281 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.213724 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.217364 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.218053 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.218158 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.218158 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.218660 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.218705 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.219085 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.221293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.221322 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.221337 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.221360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.221376 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.224503 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.224540 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.224559 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.224729 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:28.724662526 +0000 UTC m=+25.912893381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.228230 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.228266 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.228283 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.228348 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:28.728329122 +0000 UTC m=+25.916560007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.228967 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.229134 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.229548 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230582 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230645 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.229889 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.231449 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.231518 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.231542 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.229926 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230217 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230252 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230273 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230313 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.230352 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.231197 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.231831 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.232385 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.232738 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.233970 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.233222 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.233519 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.233898 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.233947 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.234153 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.234198 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.234218 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.234306 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.234828 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.236065 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.237967 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.238262 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.238722 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.238852 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.238956 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.239180 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.239428 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.239517 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.239916 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.240064 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786" exitCode=255 Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.241473 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.241922 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.243789 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.245396 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.247176 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.247212 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.251297 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.252054 4685 scope.go:117] "RemoveContainer" containerID="d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.253389 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.262344 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.266262 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.267828 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.279333 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.286579 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.286948 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.286977 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287036 4685 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287050 4685 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287062 4685 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287075 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287090 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287117 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287137 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287156 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287172 4685 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287188 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287201 4685 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287212 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287224 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287236 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287247 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287258 4685 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287269 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287281 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287292 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287303 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287315 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287326 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287337 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287348 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287360 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287371 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287383 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287395 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287433 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287460 4685 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287477 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287491 4685 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287506 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287518 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287531 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287543 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287554 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287566 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287577 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287587 4685 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287598 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287609 4685 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287621 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287634 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287647 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287659 4685 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287685 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287704 4685 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287721 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287737 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287752 4685 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287764 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287775 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287787 4685 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287801 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287812 4685 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287823 4685 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287834 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287848 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287863 4685 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287878 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287894 4685 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287910 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287926 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287941 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287953 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287965 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287977 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.287990 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288001 4685 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288013 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288025 4685 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288037 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288054 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288069 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288086 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288103 4685 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288092 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288117 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288204 4685 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288215 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288226 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288235 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288235 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288244 4685 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288275 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288288 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288302 4685 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288324 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288335 4685 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288345 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288354 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288364 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288373 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288382 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288396 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288404 4685 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288434 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288443 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288451 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288460 4685 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288468 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288476 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288486 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288496 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288507 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288517 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288532 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288542 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288554 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288564 4685 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288575 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288583 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288592 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288639 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288679 4685 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288687 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288695 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288715 4685 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288723 4685 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288731 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288739 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288748 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288757 4685 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288765 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288774 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288784 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288798 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288809 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288817 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288826 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288835 4685 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288843 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288853 4685 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288860 4685 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288869 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288878 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288886 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288895 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288909 4685 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288918 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288927 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288935 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288943 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288951 4685 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288960 4685 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288970 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288978 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288987 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.288996 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.307198 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.324638 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.324670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.324678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.324709 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.324724 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.332564 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.342538 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.378031 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.399925 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.408613 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.427529 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.427599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.427616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.427634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.427649 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: W1204 06:11:28.436844 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2589eba030d2218f8fbdfa9830b16cc3ccfda3e44f8bd27ec8815c86cca0f4b1 WatchSource:0}: Error finding container 2589eba030d2218f8fbdfa9830b16cc3ccfda3e44f8bd27ec8815c86cca0f4b1: Status 404 returned error can't find the container with id 2589eba030d2218f8fbdfa9830b16cc3ccfda3e44f8bd27ec8815c86cca0f4b1 Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.531176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.531204 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.531213 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.531226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.531236 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.634044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.634091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.634105 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.634125 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.634141 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.692355 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.693072 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:11:29.693044053 +0000 UTC m=+26.881274838 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.736223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.736279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.736291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.736310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.736324 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.794123 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.794223 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.794270 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.794304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794328 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794363 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794378 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794421 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794452 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:29.794436082 +0000 UTC m=+26.982666857 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794490 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794508 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794501 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:29.794483263 +0000 UTC m=+26.982714038 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794556 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794581 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794558 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:29.794549685 +0000 UTC m=+26.982780460 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:28 crc kubenswrapper[4685]: E1204 06:11:28.794660 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:29.794649629 +0000 UTC m=+26.982880394 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.839022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.839069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.839088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.839109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.839121 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.941057 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.941128 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.941156 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.941187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:28 crc kubenswrapper[4685]: I1204 06:11:28.941208 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:28Z","lastTransitionTime":"2025-12-04T06:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.043946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.044006 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.044027 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.044051 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.044070 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.130550 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.131665 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.134358 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.135847 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.138004 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.139135 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.140594 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.142658 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.143984 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.146187 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.147104 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.147172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.147191 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.147217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.147241 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.147355 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.149971 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.151130 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.154570 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.155764 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.157843 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.159184 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.160178 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.162500 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.163942 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.164993 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.167212 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.168165 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.170543 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.171561 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.174094 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.175761 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.176845 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.178443 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.179278 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.179826 4685 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.179939 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.181368 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.181903 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.182313 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.184675 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.186247 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.186955 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.187893 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.188880 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.189712 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.190575 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.191377 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.192204 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.192850 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.193639 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.194315 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.195295 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.197521 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.198687 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.199774 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.201390 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.203065 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.204173 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.244475 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"af60dfc9d33ad002424f4e5b748cc02e420399ab34a4ebea722b777f5345378f"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.246589 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.246615 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.246629 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ed5d385087e3bcfc84dd7fa521f84d75674fa37b551b15c607610fa3367615d4"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.248590 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.248652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.248665 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.248687 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.248702 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.249803 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.251915 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.252339 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.253353 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.253464 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2589eba030d2218f8fbdfa9830b16cc3ccfda3e44f8bd27ec8815c86cca0f4b1"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.265990 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.292812 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.306135 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.324809 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.341394 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.351131 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.351189 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.351204 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.351223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.351235 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.353172 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.364060 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.374095 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.384522 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.394544 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.414138 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.430220 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.455164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.455214 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.455226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.455246 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.455261 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.455808 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.492177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.557047 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.557099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.557108 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.557129 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.557139 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.661333 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.661463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.661486 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.661510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.661534 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.703324 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.703583 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:11:31.703540371 +0000 UTC m=+28.891771186 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.763945 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.763994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.764008 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.764028 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.764043 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.804221 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.804265 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.804287 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.804306 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804402 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804467 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:31.804453855 +0000 UTC m=+28.992684630 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804525 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804601 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804646 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804662 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804679 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:31.804649661 +0000 UTC m=+28.992880436 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804737 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:31.804710713 +0000 UTC m=+28.992941498 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804857 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804873 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804890 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:29 crc kubenswrapper[4685]: E1204 06:11:29.804945 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:31.80493294 +0000 UTC m=+28.993163715 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.866879 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.866936 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.866955 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.866979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.866998 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.969469 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.969512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.969520 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.969535 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:29 crc kubenswrapper[4685]: I1204 06:11:29.969561 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:29Z","lastTransitionTime":"2025-12-04T06:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.071637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.071714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.071738 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.071780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.071808 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.125209 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.125302 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:30 crc kubenswrapper[4685]: E1204 06:11:30.125353 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.125214 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:30 crc kubenswrapper[4685]: E1204 06:11:30.125536 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:30 crc kubenswrapper[4685]: E1204 06:11:30.125619 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.174298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.174343 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.174353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.174368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.174377 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.277814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.277868 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.277885 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.277911 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.277928 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.381834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.381901 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.381918 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.381944 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.381966 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.484838 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.484877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.484886 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.484906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.484917 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.587359 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.587468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.587492 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.587518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.587535 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.690518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.690579 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.690596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.690621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.690638 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.793194 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.793262 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.793283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.793314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.793338 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.896231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.896314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.896353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.896397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.896526 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.999359 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.999472 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.999494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:30 crc kubenswrapper[4685]: I1204 06:11:30.999922 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.000178 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:30Z","lastTransitionTime":"2025-12-04T06:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.105133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.105199 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.105215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.105238 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.105253 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.208214 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.208294 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.208319 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.208352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.208373 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.311673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.311750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.311769 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.311800 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.311818 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.415308 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.415386 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.415446 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.415479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.415503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.517375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.517447 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.517459 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.517481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.517493 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.620661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.620704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.620713 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.620728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.620739 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.721585 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.721880 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:11:35.721837343 +0000 UTC m=+32.910068168 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.722852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.722880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.722888 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.722901 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.722909 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.822993 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.823065 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.823109 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.823150 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823201 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823241 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823260 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823288 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823311 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823329 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823332 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823365 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823333 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:35.823310184 +0000 UTC m=+33.011540989 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823496 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:35.823471919 +0000 UTC m=+33.011702734 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823519 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:35.82350731 +0000 UTC m=+33.011738115 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: E1204 06:11:31.823541 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:35.823530161 +0000 UTC m=+33.011760966 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.825711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.825776 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.825801 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.825828 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.825848 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.929036 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.929107 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.929135 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.929167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:31 crc kubenswrapper[4685]: I1204 06:11:31.929185 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:31Z","lastTransitionTime":"2025-12-04T06:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.036542 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.036622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.036641 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.036666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.036684 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.125379 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.125491 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:32 crc kubenswrapper[4685]: E1204 06:11:32.125636 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:32 crc kubenswrapper[4685]: E1204 06:11:32.125850 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.126088 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:32 crc kubenswrapper[4685]: E1204 06:11:32.126398 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.138387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.138481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.138504 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.138531 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.138552 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.241305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.241345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.241357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.241373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.241386 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.262932 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.279398 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.304160 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.318251 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.331826 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.343883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.343949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.343970 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.343998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.344015 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.349726 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.362157 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.374942 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.447137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.447213 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.447237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.447267 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.447290 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.550927 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.551517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.551807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.552040 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.552263 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.655102 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.655152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.655165 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.655181 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.655191 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.758856 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.758935 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.759003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.759036 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.759058 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.774736 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.779504 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.786689 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.790297 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.805292 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.819527 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.832950 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.851969 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.861947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.861980 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.861992 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.862009 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.862022 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.867696 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.894122 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.908337 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.924250 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.939757 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.953803 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.965328 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.965384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.965396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.965439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.965456 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:32Z","lastTransitionTime":"2025-12-04T06:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.972673 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:32 crc kubenswrapper[4685]: I1204 06:11:32.988335 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:32Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.004891 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.018259 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.068682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.068720 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.068730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.068746 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.068756 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.147073 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.171086 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.171145 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.171193 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.171219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.171236 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.171235 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.189288 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.207150 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.227765 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.247483 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.270151 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.274447 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.274492 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.274505 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.274523 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.274536 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.297382 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:33Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.376805 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.376855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.376867 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.376885 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.376896 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.474806 4685 csr.go:261] certificate signing request csr-nzzcs is approved, waiting to be issued Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.479881 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.479942 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.479955 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.479972 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.479983 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.495426 4685 csr.go:257] certificate signing request csr-nzzcs is issued Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.582565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.582606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.582618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.582635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.582645 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.684769 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.684807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.684816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.684831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.684840 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.787025 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.787071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.787081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.787099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.787108 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.889155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.889209 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.889226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.889247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.889262 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.991881 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.991939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.991951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.991978 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:33 crc kubenswrapper[4685]: I1204 06:11:33.991989 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:33Z","lastTransitionTime":"2025-12-04T06:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.094977 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.095022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.095034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.095056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.095068 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.125108 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.125159 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.125108 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.125312 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.125424 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.125579 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.197621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.197665 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.197676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.197692 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.197702 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.298887 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-79fpc"] Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299280 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gsmsw"] Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299514 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-dpd9d"] Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299660 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9klkp"] Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299698 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299754 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299868 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299874 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.299991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.300008 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.300021 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.300871 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.302674 4685 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.302719 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.302819 4685 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.302860 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.302919 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.302921 4685 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.302943 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.302899 4685 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.302965 4685 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.303004 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.302966 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.303156 4685 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.303181 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.303371 4685 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.303391 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.303558 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.303698 4685 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.303735 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.305852 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.305939 4685 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 04 06:11:34 crc kubenswrapper[4685]: E1204 06:11:34.305972 4685 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.305988 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.306032 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.307803 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.318134 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.333437 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.353872 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.366120 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.376241 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.387932 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.399176 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.402100 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.402164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.402181 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.402203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.402222 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.408739 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.422450 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.434986 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445577 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/416d3991-f0c3-40a2-b997-e40e43b4ddde-proxy-tls\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445614 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-system-cni-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445637 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-os-release\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445663 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2cdd0c58-7581-4428-b193-2666a9712555-cni-binary-copy\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445705 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-cni-bin\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445726 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-conf-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445756 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/416d3991-f0c3-40a2-b997-e40e43b4ddde-mcd-auth-proxy-config\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445779 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/91fd502d-44cb-478b-930f-efcd59fb5509-hosts-file\") pod \"node-resolver-79fpc\" (UID: \"91fd502d-44cb-478b-930f-efcd59fb5509\") " pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445801 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-os-release\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445822 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-netns\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445848 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-kubelet\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445868 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-multus-certs\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445888 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-cni-binary-copy\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445908 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-cni-multus\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445932 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgqdk\" (UniqueName: \"kubernetes.io/projected/91fd502d-44cb-478b-930f-efcd59fb5509-kube-api-access-lgqdk\") pod \"node-resolver-79fpc\" (UID: \"91fd502d-44cb-478b-930f-efcd59fb5509\") " pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445965 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-cni-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.445997 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-etc-kubernetes\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446020 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/416d3991-f0c3-40a2-b997-e40e43b4ddde-rootfs\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446043 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vrr9\" (UniqueName: \"kubernetes.io/projected/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-kube-api-access-2vrr9\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446097 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vjzj\" (UniqueName: \"kubernetes.io/projected/416d3991-f0c3-40a2-b997-e40e43b4ddde-kube-api-access-8vjzj\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446144 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-k8s-cni-cncf-io\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-daemon-config\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446219 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-socket-dir-parent\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446237 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-system-cni-dir\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446257 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfjdp\" (UniqueName: \"kubernetes.io/projected/2cdd0c58-7581-4428-b193-2666a9712555-kube-api-access-hfjdp\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446276 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-cnibin\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446294 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-hostroot\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446310 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-cnibin\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446329 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.446345 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2cdd0c58-7581-4428-b193-2666a9712555-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.449033 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.462605 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.485928 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.496732 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-04 06:06:33 +0000 UTC, rotation deadline is 2026-08-26 03:41:14.084694188 +0000 UTC Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.496775 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6357h29m39.587921994s for next certificate rotation Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.502814 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.504439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.504497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.504514 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.504535 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.504547 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.524792 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.541256 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.546979 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-k8s-cni-cncf-io\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547014 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-daemon-config\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547048 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-socket-dir-parent\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547065 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-system-cni-dir\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547080 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfjdp\" (UniqueName: \"kubernetes.io/projected/2cdd0c58-7581-4428-b193-2666a9712555-kube-api-access-hfjdp\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547100 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-cnibin\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547116 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-hostroot\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547130 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-cnibin\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547149 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547168 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2cdd0c58-7581-4428-b193-2666a9712555-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547184 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/416d3991-f0c3-40a2-b997-e40e43b4ddde-proxy-tls\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547186 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-k8s-cni-cncf-io\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547213 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-socket-dir-parent\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547198 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-system-cni-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-hostroot\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547272 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-os-release\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547234 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-system-cni-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547229 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-cnibin\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547662 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2cdd0c58-7581-4428-b193-2666a9712555-cni-binary-copy\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547722 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-cni-bin\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547751 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-conf-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547773 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-cnibin\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547821 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/416d3991-f0c3-40a2-b997-e40e43b4ddde-mcd-auth-proxy-config\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547841 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-cni-bin\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547852 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/91fd502d-44cb-478b-930f-efcd59fb5509-hosts-file\") pod \"node-resolver-79fpc\" (UID: \"91fd502d-44cb-478b-930f-efcd59fb5509\") " pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547891 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-os-release\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547295 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-system-cni-dir\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547895 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-os-release\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547883 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-conf-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547944 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-netns\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547965 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-os-release\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547969 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-kubelet\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547973 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/91fd502d-44cb-478b-930f-efcd59fb5509-hosts-file\") pod \"node-resolver-79fpc\" (UID: \"91fd502d-44cb-478b-930f-efcd59fb5509\") " pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.547997 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-multus-certs\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548017 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-multus-certs\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548027 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-kubelet\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548040 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-cni-binary-copy\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-run-netns\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-cni-multus\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548132 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-host-var-lib-cni-multus\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548137 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgqdk\" (UniqueName: \"kubernetes.io/projected/91fd502d-44cb-478b-930f-efcd59fb5509-kube-api-access-lgqdk\") pod \"node-resolver-79fpc\" (UID: \"91fd502d-44cb-478b-930f-efcd59fb5509\") " pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548173 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-cni-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548194 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-etc-kubernetes\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548213 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/416d3991-f0c3-40a2-b997-e40e43b4ddde-rootfs\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548236 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vrr9\" (UniqueName: \"kubernetes.io/projected/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-kube-api-access-2vrr9\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548260 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vjzj\" (UniqueName: \"kubernetes.io/projected/416d3991-f0c3-40a2-b997-e40e43b4ddde-kube-api-access-8vjzj\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548516 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-etc-kubernetes\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548565 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/416d3991-f0c3-40a2-b997-e40e43b4ddde-rootfs\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548600 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-cni-dir\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.548706 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2cdd0c58-7581-4428-b193-2666a9712555-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.550172 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2cdd0c58-7581-4428-b193-2666a9712555-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.556824 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.567074 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.568827 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgqdk\" (UniqueName: \"kubernetes.io/projected/91fd502d-44cb-478b-930f-efcd59fb5509-kube-api-access-lgqdk\") pod \"node-resolver-79fpc\" (UID: \"91fd502d-44cb-478b-930f-efcd59fb5509\") " pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.583200 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.595332 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.607349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.607386 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.607396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.607432 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.607443 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.607673 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.611881 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-79fpc" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.620823 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: W1204 06:11:34.626563 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91fd502d_44cb_478b_930f_efcd59fb5509.slice/crio-0c98299f8df8ce5e17222971888cb581846dc60616970ef92cb8def84b9503cf WatchSource:0}: Error finding container 0c98299f8df8ce5e17222971888cb581846dc60616970ef92cb8def84b9503cf: Status 404 returned error can't find the container with id 0c98299f8df8ce5e17222971888cb581846dc60616970ef92cb8def84b9503cf Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.635057 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.681275 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n4rjr"] Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.682060 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.685046 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.685193 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.685334 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.685474 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.685564 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.685653 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.686170 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.701819 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.713493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.713530 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.713778 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.713798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.713809 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.715049 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.727232 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764614 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-script-lib\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764682 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-netns\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764729 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-config\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764778 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-kubelet\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764809 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-netd\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764836 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l97ms\" (UniqueName: \"kubernetes.io/projected/f4c7de36-abbd-448e-bf66-eecafc38068c-kube-api-access-l97ms\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764913 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-node-log\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764946 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-bin\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.764973 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-env-overrides\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765012 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-systemd-units\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765041 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-slash\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765068 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-var-lib-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-ovn-kubernetes\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765141 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4c7de36-abbd-448e-bf66-eecafc38068c-ovn-node-metrics-cert\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765184 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-systemd\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765231 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-etc-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765285 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765313 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-log-socket\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765341 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-ovn\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.765689 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.782736 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.796356 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.808738 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.816200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.816244 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.816257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.816275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.816287 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.834903 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.846126 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.861151 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.866837 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-kubelet\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.866886 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-netd\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.866912 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l97ms\" (UniqueName: \"kubernetes.io/projected/f4c7de36-abbd-448e-bf66-eecafc38068c-kube-api-access-l97ms\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.866973 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-node-log\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867000 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-bin\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867020 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-env-overrides\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867060 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-systemd-units\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867086 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-slash\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867108 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-var-lib-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867132 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-ovn-kubernetes\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867164 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4c7de36-abbd-448e-bf66-eecafc38068c-ovn-node-metrics-cert\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867205 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-systemd\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867240 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-etc-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867272 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867279 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-netd\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867314 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-node-log\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867335 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-log-socket\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867339 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-ovn-kubernetes\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867279 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-bin\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867293 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-log-socket\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867378 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-systemd-units\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867315 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-var-lib-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867429 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-etc-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867441 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867459 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-kubelet\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867462 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-slash\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867482 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-systemd\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867514 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867518 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-openvswitch\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867640 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-ovn\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-ovn\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867753 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-script-lib\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867802 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-netns\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.867826 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-config\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.868018 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-env-overrides\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.868067 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-netns\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.868654 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-script-lib\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.868737 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-config\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.874894 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4c7de36-abbd-448e-bf66-eecafc38068c-ovn-node-metrics-cert\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.874936 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.881983 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l97ms\" (UniqueName: \"kubernetes.io/projected/f4c7de36-abbd-448e-bf66-eecafc38068c-kube-api-access-l97ms\") pod \"ovnkube-node-n4rjr\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.885763 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.898796 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:34Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.918160 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.918221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.918233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.918254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:34 crc kubenswrapper[4685]: I1204 06:11:34.918266 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:34Z","lastTransitionTime":"2025-12-04T06:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.003552 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:35 crc kubenswrapper[4685]: W1204 06:11:35.019668 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4c7de36_abbd_448e_bf66_eecafc38068c.slice/crio-69c706cf67d0a5846a9f866fec57c909e9894067daa991ebecda10951dc2192e WatchSource:0}: Error finding container 69c706cf67d0a5846a9f866fec57c909e9894067daa991ebecda10951dc2192e: Status 404 returned error can't find the container with id 69c706cf67d0a5846a9f866fec57c909e9894067daa991ebecda10951dc2192e Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.020771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.020803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.020814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.020832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.020843 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.123485 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.123705 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.123910 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.124048 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.124174 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.223087 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.226255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.226313 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.226331 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.226358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.226376 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.228180 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-multus-daemon-config\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.262752 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.269058 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/416d3991-f0c3-40a2-b997-e40e43b4ddde-mcd-auth-proxy-config\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.273305 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-79fpc" event={"ID":"91fd502d-44cb-478b-930f-efcd59fb5509","Type":"ContainerStarted","Data":"67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.273494 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-79fpc" event={"ID":"91fd502d-44cb-478b-930f-efcd59fb5509","Type":"ContainerStarted","Data":"0c98299f8df8ce5e17222971888cb581846dc60616970ef92cb8def84b9503cf"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.275005 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" exitCode=0 Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.275059 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.275087 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"69c706cf67d0a5846a9f866fec57c909e9894067daa991ebecda10951dc2192e"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.289682 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.304176 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.319015 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.330543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.330818 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.330950 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.331067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.331176 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.331689 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.345672 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.359336 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.373491 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.382388 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.389469 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.398868 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vjzj\" (UniqueName: \"kubernetes.io/projected/416d3991-f0c3-40a2-b997-e40e43b4ddde-kube-api-access-8vjzj\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.418279 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.433895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.433932 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.433942 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.433959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.433969 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.456724 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.462944 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.469319 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-cni-binary-copy\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.469319 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2cdd0c58-7581-4428-b193-2666a9712555-cni-binary-copy\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.489016 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.508101 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.514131 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.518041 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.528629 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.537705 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.537747 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.537757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.537774 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.537785 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.540332 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.543522 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.548370 4685 secret.go:188] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.548505 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416d3991-f0c3-40a2-b997-e40e43b4ddde-proxy-tls podName:416d3991-f0c3-40a2-b997-e40e43b4ddde nodeName:}" failed. No retries permitted until 2025-12-04 06:11:36.048474273 +0000 UTC m=+33.236705038 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/416d3991-f0c3-40a2-b997-e40e43b4ddde-proxy-tls") pod "machine-config-daemon-gsmsw" (UID: "416d3991-f0c3-40a2-b997-e40e43b4ddde") : failed to sync secret cache: timed out waiting for the condition Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.549222 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vrr9\" (UniqueName: \"kubernetes.io/projected/736d38b0-c0ce-4db5-bee6-72ee0dc23b40-kube-api-access-2vrr9\") pod \"multus-dpd9d\" (UID: \"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\") " pod="openshift-multus/multus-dpd9d" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.552521 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.556263 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfjdp\" (UniqueName: \"kubernetes.io/projected/2cdd0c58-7581-4428-b193-2666a9712555-kube-api-access-hfjdp\") pod \"multus-additional-cni-plugins-9klkp\" (UID: \"2cdd0c58-7581-4428-b193-2666a9712555\") " pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.572184 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.580879 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.582891 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.597766 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.608787 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.623561 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.634851 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.640301 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.640355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.640368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.640384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.640395 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.645889 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.660180 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.670355 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.684279 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:35Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.691482 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.743085 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.743131 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.743142 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.743157 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.743166 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.773784 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.774026 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:11:43.773989928 +0000 UTC m=+40.962220743 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.824165 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dpd9d" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.829981 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9klkp" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.847159 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.847204 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.847216 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.847235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.847246 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:35 crc kubenswrapper[4685]: W1204 06:11:35.861470 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod736d38b0_c0ce_4db5_bee6_72ee0dc23b40.slice/crio-84655be6adf88842d7be21d2ad90c9cb3386832562ae4a145f83125a39fa578a WatchSource:0}: Error finding container 84655be6adf88842d7be21d2ad90c9cb3386832562ae4a145f83125a39fa578a: Status 404 returned error can't find the container with id 84655be6adf88842d7be21d2ad90c9cb3386832562ae4a145f83125a39fa578a Dec 04 06:11:35 crc kubenswrapper[4685]: W1204 06:11:35.861944 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cdd0c58_7581_4428_b193_2666a9712555.slice/crio-d163a4e6e5133797415fa9e2e72aebe781ee257c05ac6b69bca1f178791c5b54 WatchSource:0}: Error finding container d163a4e6e5133797415fa9e2e72aebe781ee257c05ac6b69bca1f178791c5b54: Status 404 returned error can't find the container with id d163a4e6e5133797415fa9e2e72aebe781ee257c05ac6b69bca1f178791c5b54 Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.872286 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.874953 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.874985 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.875021 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.875042 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875156 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875174 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875184 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875221 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:43.87520995 +0000 UTC m=+41.063440725 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875487 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875516 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:43.87550961 +0000 UTC m=+41.063740385 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875555 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875574 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:43.875568371 +0000 UTC m=+41.063799136 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875619 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875630 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875641 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:35 crc kubenswrapper[4685]: E1204 06:11:35.875661 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:43.875654444 +0000 UTC m=+41.063885219 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.950022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.950082 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.950095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.950114 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:35 crc kubenswrapper[4685]: I1204 06:11:35.950127 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:35Z","lastTransitionTime":"2025-12-04T06:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.052676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.052717 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.052725 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.052741 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.052750 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.076973 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/416d3991-f0c3-40a2-b997-e40e43b4ddde-proxy-tls\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.081842 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/416d3991-f0c3-40a2-b997-e40e43b4ddde-proxy-tls\") pod \"machine-config-daemon-gsmsw\" (UID: \"416d3991-f0c3-40a2-b997-e40e43b4ddde\") " pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.119932 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.125806 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.125827 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.125854 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:36 crc kubenswrapper[4685]: E1204 06:11:36.125930 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:36 crc kubenswrapper[4685]: E1204 06:11:36.126099 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:36 crc kubenswrapper[4685]: E1204 06:11:36.126184 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:36 crc kubenswrapper[4685]: W1204 06:11:36.133948 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod416d3991_f0c3_40a2_b997_e40e43b4ddde.slice/crio-42c74aec23eeaf5a711a19f6b7abca0d572b434b44c0288eac6f9c3b46f14026 WatchSource:0}: Error finding container 42c74aec23eeaf5a711a19f6b7abca0d572b434b44c0288eac6f9c3b46f14026: Status 404 returned error can't find the container with id 42c74aec23eeaf5a711a19f6b7abca0d572b434b44c0288eac6f9c3b46f14026 Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.155388 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.155447 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.155458 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.155475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.155486 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.170212 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-47k4t"] Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.170683 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.172904 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.172969 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.173102 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.173264 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.186396 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.199152 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.211403 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.225925 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.239821 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.253344 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.261671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.262613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.262704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.262811 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.263510 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.269621 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.278145 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nwgh\" (UniqueName: \"kubernetes.io/projected/043c702d-acf9-49a1-90a8-790016690f55-kube-api-access-6nwgh\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.278198 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/043c702d-acf9-49a1-90a8-790016690f55-serviceca\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.278226 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/043c702d-acf9-49a1-90a8-790016690f55-host\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.282865 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.283131 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"42c74aec23eeaf5a711a19f6b7abca0d572b434b44c0288eac6f9c3b46f14026"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.285127 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.286421 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerStarted","Data":"b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.286471 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerStarted","Data":"84655be6adf88842d7be21d2ad90c9cb3386832562ae4a145f83125a39fa578a"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.291716 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.291760 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.291771 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.291781 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.291793 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.291802 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.293154 4685 generic.go:334] "Generic (PLEG): container finished" podID="2cdd0c58-7581-4428-b193-2666a9712555" containerID="e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922" exitCode=0 Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.293251 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerDied","Data":"e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.293308 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerStarted","Data":"d163a4e6e5133797415fa9e2e72aebe781ee257c05ac6b69bca1f178791c5b54"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.299539 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.313279 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.327922 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.353393 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.365974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.366121 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.366182 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.366254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.366316 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.367323 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.379526 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/043c702d-acf9-49a1-90a8-790016690f55-serviceca\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.380132 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/043c702d-acf9-49a1-90a8-790016690f55-host\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.380730 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nwgh\" (UniqueName: \"kubernetes.io/projected/043c702d-acf9-49a1-90a8-790016690f55-kube-api-access-6nwgh\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.380393 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.380746 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/043c702d-acf9-49a1-90a8-790016690f55-serviceca\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.380370 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/043c702d-acf9-49a1-90a8-790016690f55-host\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.395599 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.399805 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nwgh\" (UniqueName: \"kubernetes.io/projected/043c702d-acf9-49a1-90a8-790016690f55-kube-api-access-6nwgh\") pod \"node-ca-47k4t\" (UID: \"043c702d-acf9-49a1-90a8-790016690f55\") " pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.409880 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.425602 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.436055 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.448969 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.463008 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.470497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.470751 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.470852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.470938 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.471033 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.477699 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.489433 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.529449 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.569656 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.573636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.573674 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.573684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.573702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.573710 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.573910 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-47k4t" Dec 04 06:11:36 crc kubenswrapper[4685]: W1204 06:11:36.586298 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod043c702d_acf9_49a1_90a8_790016690f55.slice/crio-7059094f28bd71e15d7855bed7556316cbcbb2c781a1935280afbd678597b530 WatchSource:0}: Error finding container 7059094f28bd71e15d7855bed7556316cbcbb2c781a1935280afbd678597b530: Status 404 returned error can't find the container with id 7059094f28bd71e15d7855bed7556316cbcbb2c781a1935280afbd678597b530 Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.611733 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.651806 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.678529 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.678569 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.678580 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.678597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.678606 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.690753 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.730037 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:36Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.788601 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.788666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.788685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.788711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.788729 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.891441 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.891507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.891518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.894278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.894304 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.996760 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.996834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.996855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.996903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:36 crc kubenswrapper[4685]: I1204 06:11:36.996926 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:36Z","lastTransitionTime":"2025-12-04T06:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.100475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.100513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.100545 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.100561 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.100570 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.203939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.204291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.204307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.204328 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.204340 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.298184 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.300263 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-47k4t" event={"ID":"043c702d-acf9-49a1-90a8-790016690f55","Type":"ContainerStarted","Data":"d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.300297 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-47k4t" event={"ID":"043c702d-acf9-49a1-90a8-790016690f55","Type":"ContainerStarted","Data":"7059094f28bd71e15d7855bed7556316cbcbb2c781a1935280afbd678597b530"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.304230 4685 generic.go:334] "Generic (PLEG): container finished" podID="2cdd0c58-7581-4428-b193-2666a9712555" containerID="9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc" exitCode=0 Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.304275 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerDied","Data":"9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.306134 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.306166 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.306175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.306191 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.306201 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.334132 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.380131 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.408126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.408182 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.408192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.408207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.408216 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.410720 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.422097 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.434835 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.447738 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.464766 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.474697 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.485756 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.496593 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.507550 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.509803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.509852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.509862 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.509878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.509887 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.518775 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.530495 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.540075 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.551551 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.563389 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.583731 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.593739 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.604250 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.612393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.612463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.612473 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.612489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.612498 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.614345 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.624488 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.634361 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.639324 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.639358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.639371 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.639388 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.639398 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.649250 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: E1204 06:11:37.651570 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.654599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.654633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.654644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.654660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.654672 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: E1204 06:11:37.668870 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.675629 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.675671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.675682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.675697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.675708 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: E1204 06:11:37.688167 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.691676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.691739 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.691750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.691773 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.691787 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.693535 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: E1204 06:11:37.703366 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.706458 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.706501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.706513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.706533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.706545 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: E1204 06:11:37.721000 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: E1204 06:11:37.721169 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.722839 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.722895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.722912 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.722937 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.722952 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.729956 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.773633 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.811224 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.825005 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.825062 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.825087 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.825115 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.825132 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.849937 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:37Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.927035 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.927080 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.927089 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.927104 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:37 crc kubenswrapper[4685]: I1204 06:11:37.927114 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:37Z","lastTransitionTime":"2025-12-04T06:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.030708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.030778 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.030798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.030824 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.030842 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.124635 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.124668 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.124710 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:38 crc kubenswrapper[4685]: E1204 06:11:38.124807 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:38 crc kubenswrapper[4685]: E1204 06:11:38.124883 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:38 crc kubenswrapper[4685]: E1204 06:11:38.124978 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.132843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.132901 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.132915 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.132946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.132970 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.235466 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.235510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.235524 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.235544 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.235557 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.308843 4685 generic.go:334] "Generic (PLEG): container finished" podID="2cdd0c58-7581-4428-b193-2666a9712555" containerID="bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551" exitCode=0 Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.308891 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerDied","Data":"bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.313261 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.323510 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.338568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.338610 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.338624 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.338657 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.338670 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.340235 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.353134 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.374021 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.386757 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.398018 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.416241 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.436315 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.442548 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.442577 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.442589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.442604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.442615 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.448773 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.464751 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.480738 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.495318 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.510031 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.523197 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:38Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.545899 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.545934 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.545944 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.545960 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.545970 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.649279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.649711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.649729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.649748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.649760 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.752608 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.752640 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.752647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.752662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.752671 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.855063 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.855095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.855105 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.855120 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.855130 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.957542 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.957607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.957626 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.957649 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:38 crc kubenswrapper[4685]: I1204 06:11:38.957667 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:38Z","lastTransitionTime":"2025-12-04T06:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.060663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.060729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.060749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.060776 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.060794 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.163330 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.163384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.163456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.163491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.163515 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.266991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.267067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.267099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.267125 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.267142 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.320049 4685 generic.go:334] "Generic (PLEG): container finished" podID="2cdd0c58-7581-4428-b193-2666a9712555" containerID="8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e" exitCode=0 Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.320110 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerDied","Data":"8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.370460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.370543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.370567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.370599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.370625 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.379710 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.402946 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.424629 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.443884 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.463351 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.473725 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.473841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.473861 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.473936 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.473957 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.478372 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.501927 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.519868 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.535146 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.550493 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.576264 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.576331 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.576350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.576376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.576396 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.583751 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.593047 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.605013 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.614546 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:39Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.678831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.678896 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.678914 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.678940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.678958 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.782142 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.782189 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.782206 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.782230 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.782246 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.885359 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.885512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.885545 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.885576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.885599 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.988930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.988963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.988971 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.988984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:39 crc kubenswrapper[4685]: I1204 06:11:39.988993 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:39Z","lastTransitionTime":"2025-12-04T06:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.092304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.092358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.092376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.092400 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.092457 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.124961 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.125057 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.125112 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:40 crc kubenswrapper[4685]: E1204 06:11:40.125213 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:40 crc kubenswrapper[4685]: E1204 06:11:40.125354 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:40 crc kubenswrapper[4685]: E1204 06:11:40.125582 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.196538 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.196626 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.196654 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.196685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.196710 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.298803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.299350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.299366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.299387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.299427 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.337693 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.338649 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.338711 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.338725 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.344184 4685 generic.go:334] "Generic (PLEG): container finished" podID="2cdd0c58-7581-4428-b193-2666a9712555" containerID="4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e" exitCode=0 Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.344219 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerDied","Data":"4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.357204 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.375918 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.377292 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.382597 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.394918 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.401553 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.401597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.401624 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.401644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.401656 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.405442 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.427908 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.439657 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.452177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.466245 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.479472 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.501298 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.504991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.505017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.505026 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.505053 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.505067 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.519533 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.536977 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.556088 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.568969 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.582613 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.595911 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.608340 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.608385 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.608397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.608456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.608473 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.609883 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.623187 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.637551 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.652041 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.668103 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.678177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.688754 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.707162 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.710607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.710647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.710682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.710702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.710713 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.716646 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.731476 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.745467 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.757722 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:40Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.818133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.818222 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.818247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.818280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.818316 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.920835 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.920890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.920907 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.920930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:40 crc kubenswrapper[4685]: I1204 06:11:40.920946 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:40Z","lastTransitionTime":"2025-12-04T06:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.024543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.024608 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.024627 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.024656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.024673 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.127653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.127707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.127725 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.127754 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.127772 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.231462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.231526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.231543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.231562 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.231574 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.334355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.334403 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.334443 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.334469 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.334486 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.352184 4685 generic.go:334] "Generic (PLEG): container finished" podID="2cdd0c58-7581-4428-b193-2666a9712555" containerID="4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650" exitCode=0 Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.352274 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerDied","Data":"4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.369877 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.388876 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.404056 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.418281 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.434504 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.436224 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.436259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.436271 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.436289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.436300 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.461691 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.479906 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.491664 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.505671 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.518269 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.528989 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.540342 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.541219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.541264 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.541275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.541291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.541302 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.558289 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.568940 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:41Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.643975 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.644098 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.644169 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.644204 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.644227 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.747439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.747734 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.747745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.747762 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.747775 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.850373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.850486 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.850520 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.850548 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.850571 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.953466 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.953516 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.953529 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.953548 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:41 crc kubenswrapper[4685]: I1204 06:11:41.953560 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:41Z","lastTransitionTime":"2025-12-04T06:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.056232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.056291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.056307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.056333 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.056350 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.124992 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.125028 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.125077 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:42 crc kubenswrapper[4685]: E1204 06:11:42.125112 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:42 crc kubenswrapper[4685]: E1204 06:11:42.125204 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:42 crc kubenswrapper[4685]: E1204 06:11:42.125345 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.158661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.158695 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.158704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.158720 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.158730 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.261773 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.261807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.261817 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.261830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.261838 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.361885 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" event={"ID":"2cdd0c58-7581-4428-b193-2666a9712555","Type":"ContainerStarted","Data":"bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.363623 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.363662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.363673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.363688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.363698 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.376496 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.396385 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.416939 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.429555 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.443905 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.458357 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.470300 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.483801 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.498731 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.513590 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.524596 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.540085 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.553143 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.567619 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:42Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.718380 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.718482 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.718503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.718528 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.718544 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.820799 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.820836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.820847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.820862 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.820873 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.923158 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.923196 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.923207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.923226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.923237 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:42Z","lastTransitionTime":"2025-12-04T06:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:42 crc kubenswrapper[4685]: I1204 06:11:42.968816 4685 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.025779 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.025836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.025853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.025879 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.025897 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.130217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.130258 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.130268 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.130285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.130297 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.144361 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.163024 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.181162 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.206567 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.218969 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.233368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.233438 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.233452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.233472 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.233489 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.237249 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.252885 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.271814 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.290891 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.310398 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.336686 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.337120 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.337158 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.337183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.337215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.337239 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.357605 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.371166 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/0.log" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.373326 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.376086 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b" exitCode=1 Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.376150 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.378353 4685 scope.go:117] "RemoveContainer" containerID="f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.396176 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.413888 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.429375 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.440163 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.440208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.440228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.440251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.440267 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.441040 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.454971 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.467928 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.480405 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.495681 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.495825 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.511794 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.540462 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"message\\\":\\\" 5969 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 06:11:43.019383 5969 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 06:11:43.019453 5969 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 06:11:43.019483 5969 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 06:11:43.019495 5969 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 06:11:43.019519 5969 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 06:11:43.019524 5969 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 06:11:43.019539 5969 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 06:11:43.019541 5969 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:11:43.019547 5969 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 06:11:43.019566 5969 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:11:43.019585 5969 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:11:43.019607 5969 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 06:11:43.019629 5969 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:11:43.019705 5969 factory.go:656] Stopping watch factory\\\\nI1204 06:11:43.019710 5969 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.543196 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.543239 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.543255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.543276 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.543287 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.559070 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.577919 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.595856 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.609979 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.626530 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.642222 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.646920 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.646983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.647005 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.647033 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.647052 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.665677 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"message\\\":\\\" 5969 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 06:11:43.019383 5969 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 06:11:43.019453 5969 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 06:11:43.019483 5969 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 06:11:43.019495 5969 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 06:11:43.019519 5969 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 06:11:43.019524 5969 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 06:11:43.019539 5969 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 06:11:43.019541 5969 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:11:43.019547 5969 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 06:11:43.019566 5969 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:11:43.019585 5969 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:11:43.019607 5969 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 06:11:43.019629 5969 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:11:43.019705 5969 factory.go:656] Stopping watch factory\\\\nI1204 06:11:43.019710 5969 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.681897 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.694446 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.713431 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.730080 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.749144 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.750500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.750529 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.750543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.750561 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.750573 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.762346 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.784622 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.802771 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.830601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.830845 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:11:59.830811809 +0000 UTC m=+57.019042594 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.830727 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.853119 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.853187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.853198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.853215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.853229 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.862290 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.889121 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.899510 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:43Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.931998 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.932039 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.932068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.932087 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932180 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932210 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932234 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:59.932222618 +0000 UTC m=+57.120453393 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932240 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932255 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932286 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932214 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932398 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932310 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:59.9322904 +0000 UTC m=+57.120521195 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932427 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932519 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:59.932448504 +0000 UTC m=+57.120679279 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:11:43 crc kubenswrapper[4685]: E1204 06:11:43.932573 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:59.932546827 +0000 UTC m=+57.120777672 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.955074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.955113 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.955122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.955137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:43 crc kubenswrapper[4685]: I1204 06:11:43.955145 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:43Z","lastTransitionTime":"2025-12-04T06:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.057297 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.057339 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.057348 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.057361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.057370 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.125562 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.125631 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:44 crc kubenswrapper[4685]: E1204 06:11:44.125748 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:44 crc kubenswrapper[4685]: E1204 06:11:44.125868 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.126307 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:44 crc kubenswrapper[4685]: E1204 06:11:44.126496 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.160237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.160345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.160367 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.160859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.161103 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.265152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.265211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.265260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.265283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.265817 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.369299 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.369360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.369379 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.369404 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.369451 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.382664 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/0.log" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.387369 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.389504 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.407865 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.435000 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.460998 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"message\\\":\\\" 5969 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 06:11:43.019383 5969 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 06:11:43.019453 5969 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 06:11:43.019483 5969 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 06:11:43.019495 5969 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 06:11:43.019519 5969 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 06:11:43.019524 5969 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 06:11:43.019539 5969 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 06:11:43.019541 5969 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:11:43.019547 5969 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 06:11:43.019566 5969 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:11:43.019585 5969 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:11:43.019607 5969 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 06:11:43.019629 5969 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:11:43.019705 5969 factory.go:656] Stopping watch factory\\\\nI1204 06:11:43.019710 5969 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.473260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.473299 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.473315 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.473336 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.473352 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.479119 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.495820 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.511136 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.527896 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.544721 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.557652 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.570292 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.575560 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.575616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.575633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.575659 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.575681 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.580701 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.598983 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.616965 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.635367 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:44Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.677896 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.677948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.677963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.677985 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.678002 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.780822 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.780866 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.780882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.780906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.780924 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.884272 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.884340 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.884362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.884392 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.884442 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.988014 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.988362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.988565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.988687 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:44 crc kubenswrapper[4685]: I1204 06:11:44.988782 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:44Z","lastTransitionTime":"2025-12-04T06:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.092380 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.092768 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.092895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.093030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.093159 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.196130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.196206 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.196234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.196263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.196281 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.300056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.300546 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.300714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.300895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.301040 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.393761 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/1.log" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.394844 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/0.log" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.399564 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341" exitCode=1 Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.399623 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.399705 4685 scope.go:117] "RemoveContainer" containerID="f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.400792 4685 scope.go:117] "RemoveContainer" containerID="20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341" Dec 04 06:11:45 crc kubenswrapper[4685]: E1204 06:11:45.401371 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.405764 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.405829 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.405851 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.405877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.405898 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.420834 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.440622 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.457192 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.476845 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.495165 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.508483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.508557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.508573 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.508594 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.508608 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.514552 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.545017 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48eae6298129e50dcf21685a75996155d2292854286ffd5957efa2f6093537b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"message\\\":\\\" 5969 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 06:11:43.019383 5969 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 06:11:43.019453 5969 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 06:11:43.019483 5969 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 06:11:43.019495 5969 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 06:11:43.019519 5969 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 06:11:43.019524 5969 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 06:11:43.019539 5969 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 06:11:43.019541 5969 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:11:43.019547 5969 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 06:11:43.019566 5969 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:11:43.019585 5969 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:11:43.019607 5969 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 06:11:43.019629 5969 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:11:43.019705 5969 factory.go:656] Stopping watch factory\\\\nI1204 06:11:43.019710 5969 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.558301 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.570284 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.583442 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.595531 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.611102 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.611139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.611150 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.611165 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.611174 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.612203 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.625380 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.646015 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:45Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.717079 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.717158 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.717240 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.717315 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.717373 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.821241 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.821292 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.821304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.821326 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.821340 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.924776 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.924863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.924884 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.924911 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:45 crc kubenswrapper[4685]: I1204 06:11:45.924930 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:45Z","lastTransitionTime":"2025-12-04T06:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.029372 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.029468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.029487 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.029516 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.029535 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.136291 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.136332 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.136439 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:46 crc kubenswrapper[4685]: E1204 06:11:46.136576 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:46 crc kubenswrapper[4685]: E1204 06:11:46.136978 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:46 crc kubenswrapper[4685]: E1204 06:11:46.137122 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.140451 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.140496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.140514 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.140539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.140557 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.243532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.243588 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.243608 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.243635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.243654 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.346806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.346914 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.346942 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.346974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.346996 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.407177 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/1.log" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.415243 4685 scope.go:117] "RemoveContainer" containerID="20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341" Dec 04 06:11:46 crc kubenswrapper[4685]: E1204 06:11:46.415545 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.438664 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.451067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.451117 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.451135 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.451161 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.451179 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.454696 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.466876 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.480032 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.493514 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.506714 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.515828 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.538359 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.553923 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.554017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.554029 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.554044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.554056 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.557480 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.573572 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.592024 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.609894 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.621887 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.641637 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.656976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.657040 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.657068 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.657100 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.657124 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.694929 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8"] Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.695656 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.700483 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.702246 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.717401 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.737218 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.755645 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.760042 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.760114 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.760136 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.760168 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.760186 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.761605 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68abe3c1-b9a5-4959-bec9-9609bb03b57a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.761651 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68abe3c1-b9a5-4959-bec9-9609bb03b57a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.761685 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czddt\" (UniqueName: \"kubernetes.io/projected/68abe3c1-b9a5-4959-bec9-9609bb03b57a-kube-api-access-czddt\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.761743 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68abe3c1-b9a5-4959-bec9-9609bb03b57a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.773978 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.805806 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.824396 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.836806 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.862025 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.862319 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68abe3c1-b9a5-4959-bec9-9609bb03b57a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863441 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68abe3c1-b9a5-4959-bec9-9609bb03b57a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863387 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68abe3c1-b9a5-4959-bec9-9609bb03b57a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863527 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czddt\" (UniqueName: \"kubernetes.io/projected/68abe3c1-b9a5-4959-bec9-9609bb03b57a-kube-api-access-czddt\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863815 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863888 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863562 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68abe3c1-b9a5-4959-bec9-9609bb03b57a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863911 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.863927 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.864545 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68abe3c1-b9a5-4959-bec9-9609bb03b57a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.873011 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68abe3c1-b9a5-4959-bec9-9609bb03b57a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.879985 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.891458 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czddt\" (UniqueName: \"kubernetes.io/projected/68abe3c1-b9a5-4959-bec9-9609bb03b57a-kube-api-access-czddt\") pod \"ovnkube-control-plane-749d76644c-rrbg8\" (UID: \"68abe3c1-b9a5-4959-bec9-9609bb03b57a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.895824 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.909215 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.926880 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.945657 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.961810 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.966486 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.966542 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.966559 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.966585 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.966604 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:46Z","lastTransitionTime":"2025-12-04T06:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:46 crc kubenswrapper[4685]: I1204 06:11:46.985656 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:46Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.012716 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" Dec 04 06:11:47 crc kubenswrapper[4685]: W1204 06:11:47.033361 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68abe3c1_b9a5_4959_bec9_9609bb03b57a.slice/crio-383670353c8083533cbb651a40d46ce5388a0551f8b499e939f594cfefdb90cb WatchSource:0}: Error finding container 383670353c8083533cbb651a40d46ce5388a0551f8b499e939f594cfefdb90cb: Status 404 returned error can't find the container with id 383670353c8083533cbb651a40d46ce5388a0551f8b499e939f594cfefdb90cb Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.069614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.069662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.069675 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.069699 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.069716 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.172353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.172444 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.172462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.172487 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.172503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.275107 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.275197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.275223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.275256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.275279 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.378234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.378310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.378334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.378365 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.378389 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.419366 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" event={"ID":"68abe3c1-b9a5-4959-bec9-9609bb03b57a","Type":"ContainerStarted","Data":"383670353c8083533cbb651a40d46ce5388a0551f8b499e939f594cfefdb90cb"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.486670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.486726 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.486745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.486769 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.486787 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.589111 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.589147 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.589158 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.589175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.589186 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.692446 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.692497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.692511 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.692530 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.692544 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.795812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.795880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.795902 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.795933 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.795956 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.899618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.899668 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.899687 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.899711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.899729 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.929859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.929945 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.929993 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.930023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.930045 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: E1204 06:11:47.953111 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:47Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.959162 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.959232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.959256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.959290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.959313 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:47 crc kubenswrapper[4685]: E1204 06:11:47.975501 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:47Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.985238 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.985323 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.985341 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.985367 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:47 crc kubenswrapper[4685]: I1204 06:11:47.985383 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:47Z","lastTransitionTime":"2025-12-04T06:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.004748 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.011733 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.011789 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.011804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.011825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.011838 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.036901 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.041370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.041468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.041483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.041507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.041522 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.058819 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.058989 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.060536 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.060598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.060614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.060634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.060646 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.125106 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.125132 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.125244 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.125280 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.125478 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.125585 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.163693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.163748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.163770 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.163802 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.163824 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.236959 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-55csn"] Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.237912 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.238027 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.258639 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.266017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.266049 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.266060 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.266078 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.266092 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.278117 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.278219 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrd8f\" (UniqueName: \"kubernetes.io/projected/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-kube-api-access-nrd8f\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.280615 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.294258 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.313623 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.341817 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.357616 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.369228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.369300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.369329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.369361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.369383 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.371124 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.379160 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.379274 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrd8f\" (UniqueName: \"kubernetes.io/projected/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-kube-api-access-nrd8f\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.379398 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.379519 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:48.879491887 +0000 UTC m=+46.067722692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.385481 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.401123 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrd8f\" (UniqueName: \"kubernetes.io/projected/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-kube-api-access-nrd8f\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.405753 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.425105 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" event={"ID":"68abe3c1-b9a5-4959-bec9-9609bb03b57a","Type":"ContainerStarted","Data":"d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.425148 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" event={"ID":"68abe3c1-b9a5-4959-bec9-9609bb03b57a","Type":"ContainerStarted","Data":"f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.428177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.441475 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.457186 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.471843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.471869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.471877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.471891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.471899 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.476251 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.496974 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.517006 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.535359 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.556052 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.571582 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.574470 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.574532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.574553 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.574581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.574600 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.593935 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.613920 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.631236 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.689920 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.689959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.689968 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.689986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.689997 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.690663 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.707485 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.716287 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.727960 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.741707 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.754438 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.766030 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.778708 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.792504 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.792547 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.792557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.792575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.792587 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.793597 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.804727 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.815854 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:48Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.889762 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.889972 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:48 crc kubenswrapper[4685]: E1204 06:11:48.890103 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:49.890071105 +0000 UTC m=+47.078301920 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.895684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.895736 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.895754 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.895777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.895794 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.998912 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.998988 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.999012 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.999045 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:48 crc kubenswrapper[4685]: I1204 06:11:48.999067 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:48Z","lastTransitionTime":"2025-12-04T06:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.101834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.101906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.101931 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.101960 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.101980 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.204151 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.204233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.204258 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.204287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.204306 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.306596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.306634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.306642 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.306656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.306666 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.410338 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.410630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.410640 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.410654 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.410663 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.513736 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.513810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.513833 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.513864 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.513887 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.617055 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.617129 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.617155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.617187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.617235 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.720880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.720947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.721003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.721029 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.721048 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.824857 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.824925 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.824942 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.824967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.824986 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.902102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:49 crc kubenswrapper[4685]: E1204 06:11:49.902320 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:49 crc kubenswrapper[4685]: E1204 06:11:49.902492 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:51.902452802 +0000 UTC m=+49.090683617 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.928850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.928921 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.928948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.928978 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:49 crc kubenswrapper[4685]: I1204 06:11:49.929001 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:49Z","lastTransitionTime":"2025-12-04T06:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.032508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.032580 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.032598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.032647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.032666 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.125697 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.125746 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.125748 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.125717 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:50 crc kubenswrapper[4685]: E1204 06:11:50.125913 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:50 crc kubenswrapper[4685]: E1204 06:11:50.126048 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:50 crc kubenswrapper[4685]: E1204 06:11:50.126120 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:50 crc kubenswrapper[4685]: E1204 06:11:50.126245 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.135688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.135767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.135786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.135812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.135834 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.238861 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.238950 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.238974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.239005 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.239028 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.342985 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.343046 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.343070 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.343103 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.343124 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.445802 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.445855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.445871 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.445895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.445919 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.548753 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.548836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.548858 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.548888 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.548909 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.652229 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.652307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.652328 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.652358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.652383 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.755233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.755324 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.755349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.755384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.755449 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.857811 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.857883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.857901 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.857928 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.857945 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.961280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.961352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.961370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.961394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:50 crc kubenswrapper[4685]: I1204 06:11:50.961435 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:50Z","lastTransitionTime":"2025-12-04T06:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.064069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.064132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.064149 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.064172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.064190 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.166700 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.166771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.166789 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.166813 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.166831 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.270169 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.270207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.270216 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.270231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.270239 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.373597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.373666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.373691 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.373722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.373745 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.476038 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.476097 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.476113 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.476137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.476154 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.579507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.579598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.579617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.579643 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.579662 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.682884 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.682940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.682957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.682984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.683001 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.786149 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.786226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.786252 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.786285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.786306 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.888884 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.888935 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.888948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.888966 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.888978 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.927856 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:51 crc kubenswrapper[4685]: E1204 06:11:51.928092 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:51 crc kubenswrapper[4685]: E1204 06:11:51.928263 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:11:55.92822709 +0000 UTC m=+53.116457905 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.992510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.992572 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.992589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.992612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:51 crc kubenswrapper[4685]: I1204 06:11:51.992633 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:51Z","lastTransitionTime":"2025-12-04T06:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.095492 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.095556 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.095573 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.095604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.095621 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.125013 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.125045 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.125094 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.125180 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:52 crc kubenswrapper[4685]: E1204 06:11:52.125361 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:52 crc kubenswrapper[4685]: E1204 06:11:52.125568 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:52 crc kubenswrapper[4685]: E1204 06:11:52.125658 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:52 crc kubenswrapper[4685]: E1204 06:11:52.125894 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.198293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.198346 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.198354 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.198368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.198377 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.262101 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.272849 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.283126 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.298633 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.301471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.301514 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.301525 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.301543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.301553 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.314848 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.327235 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.338835 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.356013 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.367779 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.381165 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.404365 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.404449 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.404465 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.404486 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.404501 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.430687 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.461070 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.478441 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.490298 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.501603 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.506444 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.506472 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.506483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.506498 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.506510 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.511751 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.529788 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.541365 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:52Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.609904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.609966 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.609986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.610009 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.610027 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.713357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.713464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.713488 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.713519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.713540 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.816167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.816211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.816219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.816235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.816244 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.919509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.919569 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.919593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.919622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:52 crc kubenswrapper[4685]: I1204 06:11:52.919645 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:52Z","lastTransitionTime":"2025-12-04T06:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.022251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.022312 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.022329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.022353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.022370 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.130357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.130460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.130484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.130507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.130530 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.150127 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.169630 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.188815 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.218473 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.231168 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.234526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.234602 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.234628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.234660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.234684 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.250982 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.270332 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.284951 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.301425 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.320424 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.336051 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.338046 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.338111 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.338129 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.338155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.338172 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.345797 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.354125 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.363477 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.372089 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.384709 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.395045 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:53Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.443148 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.443215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.443235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.443263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.443281 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.546207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.546269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.546283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.546307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.546323 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.649370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.649476 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.649501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.649533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.649557 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.751647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.751694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.751711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.751732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.751749 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.854504 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.854569 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.854587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.854611 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.854629 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.957003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.957067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.957078 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.957095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:53 crc kubenswrapper[4685]: I1204 06:11:53.957105 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:53Z","lastTransitionTime":"2025-12-04T06:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.059672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.059724 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.059742 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.059764 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.059781 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.125179 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.125210 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.125288 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.125302 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:54 crc kubenswrapper[4685]: E1204 06:11:54.125502 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:54 crc kubenswrapper[4685]: E1204 06:11:54.125628 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:54 crc kubenswrapper[4685]: E1204 06:11:54.125786 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:54 crc kubenswrapper[4685]: E1204 06:11:54.125903 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.162401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.162505 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.162531 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.162561 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.162585 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.265827 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.265874 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.265890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.265917 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.266241 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.370288 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.370370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.370389 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.370463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.370485 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.474386 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.474467 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.474485 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.474508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.474525 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.577391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.577488 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.577511 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.577543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.577563 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.680246 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.680311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.680327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.680349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.680382 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.783792 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.783880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.783910 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.783945 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.783970 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.886931 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.886993 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.887008 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.887029 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.887054 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.989447 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.989484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.989493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.989509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:54 crc kubenswrapper[4685]: I1204 06:11:54.989518 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:54Z","lastTransitionTime":"2025-12-04T06:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.092536 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.092606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.092630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.092663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.092689 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.195387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.195462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.195477 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.195497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.195509 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.298515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.298567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.298583 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.298604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.298620 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.401618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.401685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.401706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.401729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.401746 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.504701 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.504780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.504803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.504939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.504967 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.607614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.607651 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.607665 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.607684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.607695 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.711293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.711368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.711384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.711449 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.711467 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.814022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.814116 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.814148 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.814180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.814201 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.916715 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.916798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.916832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.916864 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.916885 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:55Z","lastTransitionTime":"2025-12-04T06:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:55 crc kubenswrapper[4685]: I1204 06:11:55.972593 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:55 crc kubenswrapper[4685]: E1204 06:11:55.972792 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:55 crc kubenswrapper[4685]: E1204 06:11:55.972915 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:03.972887189 +0000 UTC m=+61.161118004 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.019739 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.019809 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.019827 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.019854 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.019873 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.123838 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.123915 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.123939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.123968 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.123985 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.125457 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.125455 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:56 crc kubenswrapper[4685]: E1204 06:11:56.125715 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.125759 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:56 crc kubenswrapper[4685]: E1204 06:11:56.125993 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:11:56 crc kubenswrapper[4685]: E1204 06:11:56.126130 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.126335 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:56 crc kubenswrapper[4685]: E1204 06:11:56.126503 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.227214 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.227318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.227336 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.227400 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.227452 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.329774 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.329819 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.329834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.329853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.329868 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.432840 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.432904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.432923 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.432947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.432967 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.536788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.536863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.536887 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.536913 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.536932 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.639321 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.639474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.639493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.639515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.639528 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.742826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.742876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.742892 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.742908 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.742920 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.845689 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.845727 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.845736 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.845750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.845761 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.948334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.948393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.948423 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.948439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:56 crc kubenswrapper[4685]: I1204 06:11:56.948449 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:56Z","lastTransitionTime":"2025-12-04T06:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.050561 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.050607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.050619 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.050637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.050650 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.153576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.153641 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.153658 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.153681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.153700 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.256255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.256314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.256325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.256345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.256359 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.358923 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.358983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.358994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.359016 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.359028 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.461543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.461626 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.461650 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.461684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.461709 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.565933 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.566003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.566025 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.566056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.566078 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.669289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.669360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.669375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.669430 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.669450 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.773187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.773250 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.773270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.773295 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.773313 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.876153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.876497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.876528 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.876561 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.876584 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.980212 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.980273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.980291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.980313 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:57 crc kubenswrapper[4685]: I1204 06:11:57.980330 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:57Z","lastTransitionTime":"2025-12-04T06:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.083599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.083664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.083680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.083704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.083725 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.121087 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.121199 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.121225 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.121254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.121274 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.124832 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.124883 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.124895 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.125016 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.125015 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.125184 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.125234 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.125392 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.143342 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:58Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.148228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.148279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.148293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.148312 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.148330 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.171256 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:58Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.177261 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.177318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.177333 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.177359 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.177378 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.199715 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:58Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.205054 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.205094 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.205108 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.205128 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.205144 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.223348 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:58Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.227798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.227841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.227852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.227874 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.227888 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.243318 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:11:58Z is after 2025-08-24T17:21:41Z" Dec 04 06:11:58 crc kubenswrapper[4685]: E1204 06:11:58.243507 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.245558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.245633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.245655 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.245690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.245711 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.348975 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.349027 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.349038 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.349062 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.349077 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.452342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.452390 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.452400 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.452441 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.452467 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.556072 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.556134 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.556146 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.556167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.556181 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.659937 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.660052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.660089 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.660127 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.660151 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.763397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.763470 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.763483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.763503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.763515 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.866859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.866929 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.866952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.866981 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.867002 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.969442 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.969515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.969540 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.969571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:58 crc kubenswrapper[4685]: I1204 06:11:58.969594 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:58Z","lastTransitionTime":"2025-12-04T06:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.073903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.073977 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.074000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.074030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.074054 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.176915 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.176991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.177007 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.177039 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.177058 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.280702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.280770 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.280794 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.280823 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.280845 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.383537 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.383585 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.383602 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.383626 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.383643 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.486445 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.486532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.486558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.486586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.486603 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.589290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.589367 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.589390 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.589475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.589497 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.692728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.692777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.692794 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.692816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.692832 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.795520 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.795592 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.795616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.795644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.795681 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.898694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.898798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.898816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.898878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.898897 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:11:59Z","lastTransitionTime":"2025-12-04T06:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:11:59 crc kubenswrapper[4685]: I1204 06:11:59.917070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:11:59 crc kubenswrapper[4685]: E1204 06:11:59.917399 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:12:31.917376036 +0000 UTC m=+89.105606841 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.002308 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.002393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.002464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.002495 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.002516 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.018960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.019069 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.019182 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.019245 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019184 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019372 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019462 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019494 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019531 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:32.019496167 +0000 UTC m=+89.207726982 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019574 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:32.019549079 +0000 UTC m=+89.207779894 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019267 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019649 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:32.019631191 +0000 UTC m=+89.207861996 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019357 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019709 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019733 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.019795 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:32.019776656 +0000 UTC m=+89.208007471 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.105884 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.105947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.105967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.105995 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.106014 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.125537 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.125651 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.125838 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.125863 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.125955 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.126155 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.126285 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:00 crc kubenswrapper[4685]: E1204 06:12:00.126480 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.210375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.210496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.210518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.210546 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.210563 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.313974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.314030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.314041 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.314063 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.314077 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.416980 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.417087 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.417119 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.417153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.417172 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.520603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.520680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.520697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.520758 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.520778 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.623707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.623763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.623812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.623904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.623958 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.727700 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.727767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.727786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.727813 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.727829 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.830229 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.830265 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.830275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.830289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.830299 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.933637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.933705 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.933717 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.933749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:00 crc kubenswrapper[4685]: I1204 06:12:00.933765 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:00Z","lastTransitionTime":"2025-12-04T06:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.038637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.038710 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.038730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.038757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.038777 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.142030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.142082 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.142100 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.142130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.142150 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.245952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.246097 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.246123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.246156 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.246182 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.349596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.349672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.349695 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.349746 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.349765 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.452925 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.452994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.453014 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.453041 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.453097 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.555994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.556054 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.556073 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.556096 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.556114 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.660285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.660399 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.660457 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.660490 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.660508 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.763563 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.763651 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.763673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.763699 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.763715 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.868058 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.868126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.868143 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.868180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.868198 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.971377 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.971505 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.971524 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.971551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:01 crc kubenswrapper[4685]: I1204 06:12:01.971569 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:01Z","lastTransitionTime":"2025-12-04T06:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.074595 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.074662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.074685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.074716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.074735 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.125546 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.125594 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.125564 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.125561 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:02 crc kubenswrapper[4685]: E1204 06:12:02.125750 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:02 crc kubenswrapper[4685]: E1204 06:12:02.126265 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:02 crc kubenswrapper[4685]: E1204 06:12:02.126453 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:02 crc kubenswrapper[4685]: E1204 06:12:02.126626 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.126861 4685 scope.go:117] "RemoveContainer" containerID="20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.177814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.178249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.178268 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.178295 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.178316 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.281886 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.281952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.281972 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.281999 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.282026 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.384883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.384951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.384969 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.384995 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.385022 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.479226 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/1.log" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.482355 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.482983 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.487136 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.487203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.487223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.487249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.487268 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.502153 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.517500 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.532540 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.552048 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.575263 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.590273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.590335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.590352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.590375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.590391 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.600194 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.619175 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.649882 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.664834 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.676125 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.687260 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.692638 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.692673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.692683 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.692696 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.692706 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.701073 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.717377 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.727138 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.737790 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.748679 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.758038 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:02Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.795233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.795293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.795306 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.795327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.795341 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.897223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.897252 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.897263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.897278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:02 crc kubenswrapper[4685]: I1204 06:12:02.897289 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:02Z","lastTransitionTime":"2025-12-04T06:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.000232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.000293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.000305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.000325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.000336 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.105332 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.105378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.105387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.105425 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.105436 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.141188 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.164395 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.180671 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.195548 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.208482 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.208539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.208551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.208568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.208581 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.211466 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.229312 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.244552 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.257716 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.272192 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.285586 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.303873 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.311066 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.311101 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.311113 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.311130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.311143 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.320066 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.339392 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.353758 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.366077 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.378034 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.388451 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.413005 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.413060 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.413077 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.413099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.413115 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.487937 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/2.log" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.488995 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/1.log" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.496164 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7" exitCode=1 Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.496209 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.496242 4685 scope.go:117] "RemoveContainer" containerID="20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.496924 4685 scope.go:117] "RemoveContainer" containerID="d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7" Dec 04 06:12:03 crc kubenswrapper[4685]: E1204 06:12:03.497157 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.515282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.515314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.515322 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.515335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.515344 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.519899 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.535167 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.550984 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.565943 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.579831 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.590687 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.605285 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.618140 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.618168 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.618176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.618192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.618201 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.618675 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.633357 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.651059 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.663065 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.684659 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20c410d474b373d1fa3496dffb35123f144be0a916477920aec296d108452341\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:11:44Z\\\",\\\"message\\\":\\\"p:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 06:11:44.238909 6138 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239587 6138 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr in node crc\\\\nI1204 06:11:44.239598 6138 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-n4rjr after 0 failed attempt(s)\\\\nI1204 06:11:44.239608 6138 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-n4rjr\\\\nI1204 06:11:44.239303 6138 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-dpd9d\\\\nI1204 06:11:44.238929 6138 ovn.go:134] Ensuring zone local for Pod openshift-net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.696606 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.710809 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.721530 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.721573 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.721587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.721603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.721612 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.726821 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.743316 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.759394 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:03Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.826368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.826425 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.826435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.826452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.826467 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.929717 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.929778 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.929817 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.929840 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:03 crc kubenswrapper[4685]: I1204 06:12:03.929876 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:03Z","lastTransitionTime":"2025-12-04T06:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.032095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.032141 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.032154 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.032173 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.032184 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.070891 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.071107 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.071240 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:20.071204017 +0000 UTC m=+77.259434822 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.125531 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.125569 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.125579 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.125683 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.125700 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.125896 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.125927 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.125973 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.136234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.136273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.136283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.136300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.136312 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.239387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.239490 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.239513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.239543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.239563 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.342600 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.342660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.342681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.342707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.342730 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.445509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.445572 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.445590 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.445615 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.445633 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.501594 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/2.log" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.506198 4685 scope.go:117] "RemoveContainer" containerID="d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7" Dec 04 06:12:04 crc kubenswrapper[4685]: E1204 06:12:04.506502 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.520475 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.532693 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.546158 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.549622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.549692 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.549715 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.549744 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.549836 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.562177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.582951 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.596956 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.613004 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.628096 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.653027 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.653077 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.653091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.653113 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.653128 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.655196 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.667495 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.686936 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.705614 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.721572 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.737945 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.755862 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.755922 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.755944 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.755971 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.755991 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.756143 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.785057 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.799894 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:04Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.859320 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.859368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.859378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.859397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.859427 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.962474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.962518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.962528 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.962542 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:04 crc kubenswrapper[4685]: I1204 06:12:04.962552 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:04Z","lastTransitionTime":"2025-12-04T06:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.064921 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.064976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.064998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.065065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.065087 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.168456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.168516 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.168538 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.168568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.168587 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.271965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.272027 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.272050 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.272077 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.272098 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.374947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.374986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.375001 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.375023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.375037 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.477277 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.477304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.477311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.477323 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.477331 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.579921 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.580182 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.580192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.580205 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.580214 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.683631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.683699 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.683722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.683751 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.683772 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.786670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.786727 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.786749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.786778 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.786800 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.889936 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.889994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.890013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.890036 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.890054 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.992551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.992643 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.992661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.992690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:05 crc kubenswrapper[4685]: I1204 06:12:05.992708 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:05Z","lastTransitionTime":"2025-12-04T06:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.095450 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.095517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.095533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.095554 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.095569 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.125023 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.125062 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:06 crc kubenswrapper[4685]: E1204 06:12:06.125192 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.125201 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.125246 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:06 crc kubenswrapper[4685]: E1204 06:12:06.125401 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:06 crc kubenswrapper[4685]: E1204 06:12:06.125514 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:06 crc kubenswrapper[4685]: E1204 06:12:06.125606 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.198698 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.198737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.198749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.198766 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.198777 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.301231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.301275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.301291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.301314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.301331 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.403448 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.403486 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.403496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.403512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.403524 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.505587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.505624 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.505632 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.505646 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.505655 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.608471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.608532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.608553 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.608577 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.608614 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.714549 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.714637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.714652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.714675 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.714751 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.816854 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.816883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.816890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.816903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.816913 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.918808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.918842 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.918853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.918870 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:06 crc kubenswrapper[4685]: I1204 06:12:06.918881 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:06Z","lastTransitionTime":"2025-12-04T06:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.021300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.021342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.021352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.021368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.021379 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.123435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.123467 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.123478 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.123493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.123503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.225876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.225921 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.225934 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.225951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.225963 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.328740 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.328790 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.328804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.328821 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.328830 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.431568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.431603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.431611 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.431626 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.431637 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.534148 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.534227 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.534245 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.534270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.534306 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.636612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.636663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.636678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.636700 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.636716 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.738937 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.738995 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.739019 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.739049 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.739069 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.841986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.842022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.842032 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.842045 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.842056 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.944909 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.944957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.944965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.944985 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:07 crc kubenswrapper[4685]: I1204 06:12:07.944998 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:07Z","lastTransitionTime":"2025-12-04T06:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.047243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.047514 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.047565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.047603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.047635 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.124905 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.125021 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.125126 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.125180 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.125242 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.125301 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.125478 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.125533 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.150042 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.150102 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.150126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.150155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.150178 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.252756 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.252807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.252818 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.252839 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.252853 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.263174 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.263230 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.263249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.263273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.263292 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.278610 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:08Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.282479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.282523 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.282535 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.282555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.282567 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.292715 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:08Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.296069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.296103 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.296111 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.296125 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.296135 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.307703 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:08Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.310917 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.310948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.310958 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.310974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.310985 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.359547 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:08Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.362305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.362345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.362355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.362373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.362384 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.373023 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:08Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:08 crc kubenswrapper[4685]: E1204 06:12:08.373149 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.375085 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.375240 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.375278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.375329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.375355 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.477488 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.477546 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.477562 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.477587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.477614 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.579950 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.579998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.580010 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.580024 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.580035 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.682500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.682554 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.682568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.682585 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.682596 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.785557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.785849 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.785958 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.786034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.786050 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.889025 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.889081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.889099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.889122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.889140 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.991280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.991323 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.991337 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.991353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:08 crc kubenswrapper[4685]: I1204 06:12:08.991362 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:08Z","lastTransitionTime":"2025-12-04T06:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.094197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.094254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.094266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.094285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.094297 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.196633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.196674 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.196684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.196699 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.196709 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.298913 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.298946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.298954 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.298965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.298973 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.401016 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.401052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.401059 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.401073 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.401080 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.504906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.504946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.504956 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.504972 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.504982 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.608371 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.608460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.608478 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.608501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.608519 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.711104 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.711162 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.711175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.711195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.711208 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.814136 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.814198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.814211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.814232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.814244 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.916684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.916741 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.916757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.916782 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:09 crc kubenswrapper[4685]: I1204 06:12:09.916799 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:09Z","lastTransitionTime":"2025-12-04T06:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.019503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.019544 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.019560 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.019586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.019604 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.121827 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.121891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.121913 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.121943 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.121959 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.125090 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.125110 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.125145 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:10 crc kubenswrapper[4685]: E1204 06:12:10.125188 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.125225 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:10 crc kubenswrapper[4685]: E1204 06:12:10.125334 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:10 crc kubenswrapper[4685]: E1204 06:12:10.125368 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:10 crc kubenswrapper[4685]: E1204 06:12:10.125530 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.224994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.225035 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.225046 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.225061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.225072 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.327601 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.327676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.327690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.327706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.327716 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.429835 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.429887 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.429904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.429927 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.429943 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.533108 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.533163 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.533183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.533208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.533228 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.636040 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.636113 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.636139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.636171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.636195 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.738921 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.738982 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.738998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.739022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.739041 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.842195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.842266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.842287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.842314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.842332 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.944918 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.944972 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.944993 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.945018 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:10 crc kubenswrapper[4685]: I1204 06:12:10.945035 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:10Z","lastTransitionTime":"2025-12-04T06:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.054310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.054361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.054377 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.054401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.054462 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.157117 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.157181 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.157203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.157242 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.157261 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.260318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.260441 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.260558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.260593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.260615 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.364166 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.364219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.364237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.364265 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.364289 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.467241 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.467290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.467298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.467312 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.467320 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.570282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.570382 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.570469 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.570526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.570589 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.673790 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.673855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.673872 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.673896 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.673914 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.776041 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.776115 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.776136 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.776164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.776185 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.878426 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.878471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.878479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.878494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.878503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.981051 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.981079 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.981088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.981116 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:11 crc kubenswrapper[4685]: I1204 06:12:11.981127 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:11Z","lastTransitionTime":"2025-12-04T06:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.083977 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.084313 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.084464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.084592 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.084750 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.125370 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:12 crc kubenswrapper[4685]: E1204 06:12:12.125611 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.125390 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:12 crc kubenswrapper[4685]: E1204 06:12:12.126170 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.125665 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.125639 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:12 crc kubenswrapper[4685]: E1204 06:12:12.126742 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:12 crc kubenswrapper[4685]: E1204 06:12:12.126935 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.186445 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.186682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.186750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.186886 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.186971 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.289692 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.289766 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.289789 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.289820 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.289843 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.392509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.392547 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.392557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.392571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.392580 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.494964 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.495254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.495471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.495558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.495643 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.599195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.599257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.599274 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.599299 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.599318 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.701973 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.702040 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.702060 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.702085 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.702103 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.803979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.804016 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.804028 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.804047 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.804059 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.906562 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.906621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.906637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.906664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:12 crc kubenswrapper[4685]: I1204 06:12:12.906681 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:12Z","lastTransitionTime":"2025-12-04T06:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.009578 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.009843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.009959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.010047 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.010144 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.112503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.112566 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.112590 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.112616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.112640 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.141129 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.156577 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.168836 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.185115 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.200749 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.214495 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.215397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.215471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.215488 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.215512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.215532 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.231399 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.246979 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.265773 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.276316 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.286028 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.295624 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.304917 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.316832 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.318866 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.318896 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.318905 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.318953 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.318966 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.328850 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.341796 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.352391 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:13Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.421378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.421468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.421480 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.421498 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.421509 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.523822 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.524316 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.524346 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.524378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.524403 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.626652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.626706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.626716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.626734 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.626744 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.729240 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.729287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.729304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.729327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.729343 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.831462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.831506 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.831517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.831534 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.831544 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.934159 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.934213 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.934229 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.934252 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:13 crc kubenswrapper[4685]: I1204 06:12:13.934269 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:13Z","lastTransitionTime":"2025-12-04T06:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.037402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.037479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.037491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.037526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.037538 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.124644 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.124704 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.124716 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.124734 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:14 crc kubenswrapper[4685]: E1204 06:12:14.124833 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:14 crc kubenswrapper[4685]: E1204 06:12:14.124958 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:14 crc kubenswrapper[4685]: E1204 06:12:14.125076 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:14 crc kubenswrapper[4685]: E1204 06:12:14.125170 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.140345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.140394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.140420 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.140439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.140452 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.243036 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.243080 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.243090 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.243103 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.243111 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.345219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.345264 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.345272 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.345284 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.345294 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.446986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.447042 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.447055 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.447078 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.447094 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.549079 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.549136 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.549153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.549175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.549193 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.651021 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.651057 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.651067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.651083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.651093 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.753545 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.753612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.753635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.753665 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.753695 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.856531 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.856592 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.856609 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.856633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.856650 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.965217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.965260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.965269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.965283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:14 crc kubenswrapper[4685]: I1204 06:12:14.965293 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:14Z","lastTransitionTime":"2025-12-04T06:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.067582 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.067642 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.067654 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.067674 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.067686 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.170777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.170832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.170843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.170862 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.170877 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.274234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.274278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.274289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.274306 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.274316 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.377391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.377471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.377481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.377501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.377516 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.480723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.480784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.480794 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.480817 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.480831 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.583512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.583574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.583587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.583614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.583627 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.686385 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.686479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.686493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.686524 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.686542 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.788787 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.788840 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.788859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.788884 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.788901 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.891708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.891753 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.891762 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.891777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.891786 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.994840 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.994870 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.994879 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.994893 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:15 crc kubenswrapper[4685]: I1204 06:12:15.994903 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:15Z","lastTransitionTime":"2025-12-04T06:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.096737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.096783 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.096795 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.096814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.096827 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.124948 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.125065 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.125062 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.124967 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:16 crc kubenswrapper[4685]: E1204 06:12:16.125160 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:16 crc kubenswrapper[4685]: E1204 06:12:16.125261 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:16 crc kubenswrapper[4685]: E1204 06:12:16.125506 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:16 crc kubenswrapper[4685]: E1204 06:12:16.125648 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.199836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.199916 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.199929 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.199952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.199966 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.302427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.302468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.302476 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.302491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.302500 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.405056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.405099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.405110 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.405127 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.405135 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.508579 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.508671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.508696 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.508731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.508760 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.611177 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.611237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.611256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.611280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.611297 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.713867 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.713930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.713948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.713972 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.713989 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.816184 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.816247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.816264 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.816287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.816306 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.919185 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.919275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.919300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.919330 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:16 crc kubenswrapper[4685]: I1204 06:12:16.919350 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:16Z","lastTransitionTime":"2025-12-04T06:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.025081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.025145 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.025166 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.025198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.025221 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.127360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.127432 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.127443 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.127459 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.127470 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.229951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.229986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.229997 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.230012 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.230024 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.333093 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.333191 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.333209 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.333264 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.333298 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.436445 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.436504 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.436523 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.436552 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.436570 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.539682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.539753 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.539781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.539814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.539839 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.643074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.643147 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.643171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.643200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.643224 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.746617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.746678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.746697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.746721 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.746740 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.849821 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.849878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.849895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.849922 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.849939 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.953050 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.953131 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.953167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.953198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:17 crc kubenswrapper[4685]: I1204 06:12:17.953218 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:17Z","lastTransitionTime":"2025-12-04T06:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.055604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.055660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.055677 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.055702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.055719 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.124656 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.124726 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.124732 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.124656 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.124867 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.125091 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.125222 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.125388 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.160125 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.160191 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.160209 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.160238 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.160256 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.262438 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.262475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.262485 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.262500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.262511 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.366000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.366065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.366082 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.366108 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.366124 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.469631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.469692 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.469708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.469732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.469749 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.573218 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.573284 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.573302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.573327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.573344 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.677061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.677135 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.677159 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.677190 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.677394 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.721852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.721903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.721914 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.721930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.721941 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.744477 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:18Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.750355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.750517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.750575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.750603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.750667 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.766659 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:18Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.771581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.771681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.771706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.771772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.771789 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.794074 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:18Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.799103 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.799255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.799281 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.799302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.799318 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.817452 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:18Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.821909 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.821976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.821999 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.822031 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.822052 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.838473 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:18Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:18 crc kubenswrapper[4685]: E1204 06:12:18.838714 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.840884 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.840935 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.840954 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.840978 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.840998 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.944212 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.944274 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.944293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.944318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:18 crc kubenswrapper[4685]: I1204 06:12:18.944340 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:18Z","lastTransitionTime":"2025-12-04T06:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.047597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.047642 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.047662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.047685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.047703 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.130576 4685 scope.go:117] "RemoveContainer" containerID="d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7" Dec 04 06:12:19 crc kubenswrapper[4685]: E1204 06:12:19.131069 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.150076 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.150135 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.150153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.150176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.150193 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.252771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.252843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.252869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.252898 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.252922 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.356576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.356648 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.356671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.356701 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.356723 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.460186 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.460260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.460282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.460311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.460333 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.563375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.563489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.563510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.563534 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.563552 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.666243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.666307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.666324 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.666349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.666366 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.769135 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.769210 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.769232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.769263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.769285 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.872450 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.872525 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.872551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.872586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.872611 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.977152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.977215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.977249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.977280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:19 crc kubenswrapper[4685]: I1204 06:12:19.977302 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:19Z","lastTransitionTime":"2025-12-04T06:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.080350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.080477 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.080518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.080560 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.080582 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.125083 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.125083 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:20 crc kubenswrapper[4685]: E1204 06:12:20.125295 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.125123 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:20 crc kubenswrapper[4685]: E1204 06:12:20.125326 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.125113 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:20 crc kubenswrapper[4685]: E1204 06:12:20.125554 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:20 crc kubenswrapper[4685]: E1204 06:12:20.125702 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.152010 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:20 crc kubenswrapper[4685]: E1204 06:12:20.152217 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:12:20 crc kubenswrapper[4685]: E1204 06:12:20.152310 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:12:52.152280514 +0000 UTC m=+109.340511329 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.183123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.183169 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.183183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.183198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.183209 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.285752 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.285810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.285832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.285861 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.285882 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.389321 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.389375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.389393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.389448 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.389465 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.492235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.492303 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.492322 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.492346 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.492362 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.596257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.596317 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.596334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.596356 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.596373 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.699368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.699452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.699464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.699482 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.699494 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.801718 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.801750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.801758 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.801774 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.801786 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.904555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.904594 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.904608 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.904628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:20 crc kubenswrapper[4685]: I1204 06:12:20.904641 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:20Z","lastTransitionTime":"2025-12-04T06:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.008167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.008235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.008259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.008291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.008313 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.110871 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.110908 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.110917 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.110930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.110954 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.213995 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.214060 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.214076 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.214095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.214107 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.317138 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.317208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.317231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.317262 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.317283 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.420588 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.420647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.420658 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.420674 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.420684 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.523071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.523114 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.523129 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.523148 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.523160 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.560937 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/0.log" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.560991 4685 generic.go:334] "Generic (PLEG): container finished" podID="736d38b0-c0ce-4db5-bee6-72ee0dc23b40" containerID="b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3" exitCode=1 Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.561017 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerDied","Data":"b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.561358 4685 scope.go:117] "RemoveContainer" containerID="b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.576541 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.593586 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.610753 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.625314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.625362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.625376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.625394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.625430 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.635402 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:21Z\\\",\\\"message\\\":\\\"2025-12-04T06:11:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd\\\\n2025-12-04T06:11:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd to /host/opt/cni/bin/\\\\n2025-12-04T06:11:36Z [verbose] multus-daemon started\\\\n2025-12-04T06:11:36Z [verbose] Readiness Indicator file check\\\\n2025-12-04T06:12:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.657815 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.672354 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.690171 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.703774 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.718348 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.728334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.728371 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.728385 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.728401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.728427 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.732752 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.748986 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.762970 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.777391 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.798011 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.814458 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.824709 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.831285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.831348 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.831357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.831372 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.831380 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.837986 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:21Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.934240 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.934283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.934295 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.934314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:21 crc kubenswrapper[4685]: I1204 06:12:21.934325 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:21Z","lastTransitionTime":"2025-12-04T06:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.037483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.037551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.037582 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.037609 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.037627 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.124660 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:22 crc kubenswrapper[4685]: E1204 06:12:22.124830 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.125107 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:22 crc kubenswrapper[4685]: E1204 06:12:22.125222 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.125696 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:22 crc kubenswrapper[4685]: E1204 06:12:22.125798 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.125900 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:22 crc kubenswrapper[4685]: E1204 06:12:22.125979 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.145311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.145381 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.145404 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.145468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.145489 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.251253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.251307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.251325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.251348 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.251366 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.354763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.354796 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.354804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.354818 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.354833 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.457221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.457268 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.457279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.457294 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.457305 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.561245 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.561620 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.561635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.561652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.561662 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.566297 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/0.log" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.566370 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerStarted","Data":"b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.589316 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.609353 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:21Z\\\",\\\"message\\\":\\\"2025-12-04T06:11:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd\\\\n2025-12-04T06:11:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd to /host/opt/cni/bin/\\\\n2025-12-04T06:11:36Z [verbose] multus-daemon started\\\\n2025-12-04T06:11:36Z [verbose] Readiness Indicator file check\\\\n2025-12-04T06:12:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.640270 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.655161 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.664023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.664082 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.664099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.664123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.664146 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.669976 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.685803 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.703446 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.723664 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.763550 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.766188 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.766237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.766255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.766277 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.766293 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.781827 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.796079 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.806982 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.819723 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.828746 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.846251 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.862292 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.868780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.868820 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.868831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.868847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.868861 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.877866 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:22Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.971963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.972017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.972034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.972059 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:22 crc kubenswrapper[4685]: I1204 06:12:22.972076 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:22Z","lastTransitionTime":"2025-12-04T06:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.075195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.075251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.075285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.075303 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.075315 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.151884 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.171552 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.178291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.178352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.178369 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.178396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.178443 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.189169 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.213163 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.229514 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.246790 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.266303 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.279995 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.281744 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.281795 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.281811 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.281833 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.281849 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.293843 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.313629 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:21Z\\\",\\\"message\\\":\\\"2025-12-04T06:11:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd\\\\n2025-12-04T06:11:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd to /host/opt/cni/bin/\\\\n2025-12-04T06:11:36Z [verbose] multus-daemon started\\\\n2025-12-04T06:11:36Z [verbose] Readiness Indicator file check\\\\n2025-12-04T06:12:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.342706 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.354157 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.373708 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.384264 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.384312 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.384369 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.384394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.384473 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.393156 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.408530 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.420538 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.434179 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:23Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.486963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.487237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.487247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.487261 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.487270 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.589568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.589660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.589685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.589714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.589731 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.693495 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.693567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.693587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.693612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.693629 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.796622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.796716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.796727 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.796748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.796760 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.899172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.899233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.899246 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.899270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:23 crc kubenswrapper[4685]: I1204 06:12:23.899297 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:23Z","lastTransitionTime":"2025-12-04T06:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.002319 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.002384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.002401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.002460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.002479 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.105284 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.105351 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.105373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.105399 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.105473 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.124739 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.124805 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.124775 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.124745 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:24 crc kubenswrapper[4685]: E1204 06:12:24.125002 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:24 crc kubenswrapper[4685]: E1204 06:12:24.125129 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:24 crc kubenswrapper[4685]: E1204 06:12:24.125260 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:24 crc kubenswrapper[4685]: E1204 06:12:24.125398 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.209013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.209094 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.209115 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.209146 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.209169 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.312568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.312653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.312665 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.312707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.312723 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.416501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.416604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.416625 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.416652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.416674 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.519115 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.519152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.519161 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.519175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.519188 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.621501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.621557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.621574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.621598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.621617 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.725831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.725903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.725925 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.725949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.725965 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.828705 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.828775 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.828800 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.828826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.828841 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.931608 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.931755 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.931780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.931851 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:24 crc kubenswrapper[4685]: I1204 06:12:24.931872 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:24Z","lastTransitionTime":"2025-12-04T06:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.034740 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.034800 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.034822 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.034853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.034874 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.137883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.137949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.137971 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.137998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.138019 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.240598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.240701 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.240727 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.240752 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.240770 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.342784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.342833 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.342849 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.342872 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.342888 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.445721 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.445790 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.445814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.445841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.445862 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.548588 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.548644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.548682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.548724 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.548747 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.651396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.651489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.651511 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.651535 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.651551 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.754220 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.754302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.754325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.754352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.754374 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.856876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.856929 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.856947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.856970 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.856985 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.959969 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.960025 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.960041 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.960067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:25 crc kubenswrapper[4685]: I1204 06:12:25.960084 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:25Z","lastTransitionTime":"2025-12-04T06:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.063088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.063138 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.063157 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.063180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.063213 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.124968 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.125024 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.125068 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.124984 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:26 crc kubenswrapper[4685]: E1204 06:12:26.125167 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:26 crc kubenswrapper[4685]: E1204 06:12:26.125290 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:26 crc kubenswrapper[4685]: E1204 06:12:26.125406 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:26 crc kubenswrapper[4685]: E1204 06:12:26.125537 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.166544 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.166634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.166656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.166681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.166699 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.270253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.270322 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.270345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.270375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.270395 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.373584 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.373660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.373677 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.374105 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.374164 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.477925 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.477983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.478000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.478024 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.478042 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.579798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.579850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.579870 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.579894 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.579910 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.683153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.683215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.683232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.683259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.683278 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.785727 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.785791 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.785808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.785831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.785848 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.888175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.888219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.888228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.888243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.888252 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.991436 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.991499 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.991511 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.991532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:26 crc kubenswrapper[4685]: I1204 06:12:26.991547 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:26Z","lastTransitionTime":"2025-12-04T06:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.094394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.094494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.094513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.094539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.094557 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.199464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.199539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.199565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.199596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.199619 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.303698 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.303779 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.303803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.303832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.303854 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.407117 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.407180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.407197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.407221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.407242 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.510266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.510325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.510343 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.510366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.510383 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.614004 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.614069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.614086 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.614111 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.614129 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.717981 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.718045 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.718067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.718093 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.718115 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.821616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.821695 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.821718 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.821741 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.821759 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.924498 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.924574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.924589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.924608 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:27 crc kubenswrapper[4685]: I1204 06:12:27.924619 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:27Z","lastTransitionTime":"2025-12-04T06:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.027323 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.027384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.027402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.027461 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.027482 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.124691 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.124790 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.124858 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.125112 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:28 crc kubenswrapper[4685]: E1204 06:12:28.125091 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:28 crc kubenswrapper[4685]: E1204 06:12:28.125248 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:28 crc kubenswrapper[4685]: E1204 06:12:28.125314 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:28 crc kubenswrapper[4685]: E1204 06:12:28.125478 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.130378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.130400 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.130428 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.130446 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.130455 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.233305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.233804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.233973 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.234150 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.234341 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.337763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.337835 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.337857 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.337883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.337901 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.441982 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.442442 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.442618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.442772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.442905 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.546283 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.546350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.546371 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.546401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.546448 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.649219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.649287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.649306 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.649335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.649353 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.752668 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.752748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.752786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.752825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.752848 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.855449 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.855489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.855503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.855523 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.855537 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.958444 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.958501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.958517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.958541 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:28 crc kubenswrapper[4685]: I1204 06:12:28.958558 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:28Z","lastTransitionTime":"2025-12-04T06:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.061758 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.061883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.061917 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.061946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.061966 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.164088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.164173 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.164195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.164227 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.164252 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.182653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.182704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.182721 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.182742 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.182758 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: E1204 06:12:29.200388 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.204917 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.204952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.204962 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.204977 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.204986 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: E1204 06:12:29.222035 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.226737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.226776 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.226784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.226799 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.226814 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: E1204 06:12:29.243274 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.252380 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.252487 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.252507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.252539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.252564 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: E1204 06:12:29.271148 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.275582 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.275642 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.275660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.275684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.275700 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: E1204 06:12:29.293517 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"19f8f547-1e23-412e-9678-012276047599\\\",\\\"systemUUID\\\":\\\"83da537d-e4a5-4b93-8404-e36573fe734d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:29Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:29 crc kubenswrapper[4685]: E1204 06:12:29.293737 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.296114 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.296161 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.296171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.296188 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.296197 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.399601 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.399663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.399681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.399709 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.399728 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.504503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.504604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.504662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.504690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.504744 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.608112 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.608186 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.608205 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.608229 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.608249 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.711503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.711562 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.711580 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.711605 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.711623 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.814617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.814693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.814717 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.814745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.814767 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.918350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.918456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.918475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.918500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:29 crc kubenswrapper[4685]: I1204 06:12:29.918522 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:29Z","lastTransitionTime":"2025-12-04T06:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.021629 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.021707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.021729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.021758 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.021780 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124861 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124881 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124878 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124878 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.124873 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: E1204 06:12:30.125038 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.125113 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: E1204 06:12:30.125197 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:30 crc kubenswrapper[4685]: E1204 06:12:30.125356 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:30 crc kubenswrapper[4685]: E1204 06:12:30.125501 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.228071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.228131 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.228149 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.228175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.228192 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.331185 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.331270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.331294 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.331325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.331350 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.434712 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.434788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.434811 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.434843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.434906 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.538125 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.538195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.538218 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.538249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.538273 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.642149 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.642236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.642304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.642349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.642368 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.746126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.746189 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.746208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.746260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.746280 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.850016 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.850179 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.850208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.850234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.850251 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.953987 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.954061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.954081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.954109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:30 crc kubenswrapper[4685]: I1204 06:12:30.954129 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:30Z","lastTransitionTime":"2025-12-04T06:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.056180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.056229 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.056241 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.056260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.056272 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.125871 4685 scope.go:117] "RemoveContainer" containerID="d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.158887 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.158946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.158961 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.158984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.159000 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.263023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.263101 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.263124 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.263152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.263175 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.366448 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.366564 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.366629 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.366656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.366674 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.469429 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.469466 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.469475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.469492 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.469503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.572591 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.572666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.572688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.572717 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.572760 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.600958 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/2.log" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.607577 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.609148 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.630915 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3016ebce-315b-4e25-a107-8bb9b5ead632\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5d484d1a86b4e87980ae2d5726480d3cfae95b70ef80c3b7c2f1fb0bd3fb37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86f9ef655c98a9b3c9ef07b97f6bbc6c7b9bce1105ce15ec718277a5d43d83bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3002d2ab0336633be30ddbc1edb40b936fa07280bb7b156f3ec4a3778f216de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.651466 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc7e1a84-7b7e-4c7d-8d0f-516b523c061c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5a76e6aad76444a49f99541eddc88d418c77acac3cb9fc5b9ac4de37d677c81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa9fefca9032918366b5a11c23aa31666167dfe363b532877dd0959791e205a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1153a9bd59f172156a6156f15f181acea0772a18d8d2f22036d3f722affcb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c7b3fb0dd50fa7ed274411eac5f8d91bfdd4e5e59455654423c1252eb1a0aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.667851 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.675397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.675463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.675479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.675506 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.675523 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.690060 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dpd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736d38b0-c0ce-4db5-bee6-72ee0dc23b40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:12:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:21Z\\\",\\\"message\\\":\\\"2025-12-04T06:11:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd\\\\n2025-12-04T06:11:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d0966bc3-4aa7-4a88-ab96-2a696b7c3bcd to /host/opt/cni/bin/\\\\n2025-12-04T06:11:36Z [verbose] multus-daemon started\\\\n2025-12-04T06:11:36Z [verbose] Readiness Indicator file check\\\\n2025-12-04T06:12:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vrr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dpd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.722093 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c7de36-abbd-448e-bf66-eecafc38068c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T06:12:03Z\\\",\\\"message\\\":\\\"pping watch factory\\\\nI1204 06:12:03.102697 6354 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 06:12:03.102708 6354 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 06:12:03.102722 6354 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 06:12:03.102728 6354 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102732 6354 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 06:12:03.102810 6354 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.102879 6354 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103079 6354 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103144 6354 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103175 6354 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 06:12:03.103265 6354 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:12:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l97ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n4rjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.741396 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-47k4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"043c702d-acf9-49a1-90a8-790016690f55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d204ffb0092080a010e2f3c32d66442141882925a9945624967f4e113cd75dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-47k4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.757433 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.768973 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68abe3c1-b9a5-4959-bec9-9609bb03b57a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f34de1d563756ec96179f81ab97cdc50684587e3ca1ce2be9d19b5ad51db6750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cad8df0914bc8fa9c46e05702f13ce1b1b0c0879a6e7032fd2c8a6dc92c128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czddt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrbg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.777877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.777914 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.777925 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.777941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.777955 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.780626 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55csn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrd8f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55csn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.793801 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"546be9fc-d625-4c4b-a8e1-9fc0ff588b2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T06:11:27Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 06:11:15.418750 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 06:11:15.420490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-989261516/tls.crt::/tmp/serving-cert-989261516/tls.key\\\\\\\"\\\\nI1204 06:11:27.159644 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 06:11:27.169558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 06:11:27.169591 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 06:11:27.169618 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 06:11:27.169623 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 06:11:27.178061 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 06:11:27.178110 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 06:11:27.178122 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1204 06:11:27.178118 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1204 06:11:27.178135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 06:11:27.178150 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 06:11:27.178158 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 06:11:27.178166 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1204 06:11:27.180317 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.807191 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839434387f1ba1e80543c4f4c2052eea46e18bc06daa72412dc44cc7f5540040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.824772 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9b7cffff90dff2d62bad92737e83f97848e7d7400d2b47ff8d1eb96257ef1cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.836617 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416d3991-f0c3-40a2-b997-e40e43b4ddde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02e46db120efc407ed7a736d004c24b8cda6fd0966a0ea1cd4355dcedac4123b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vjzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gsmsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.850250 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.861799 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e57cf1f4fd3108eb0735c0d6a6b0de5bbb68bf2474b8dc8fbd2f83ddfdb3963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbd25b1fd0f9fed72575a1758ebe9e2189e042cd5b047df98261c3dd32b100ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.873481 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-79fpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91fd502d-44cb-478b-930f-efcd59fb5509\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67dae765682162b313c81495c2956fc637e2f5ecd02d64d5cda4146669ac5b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgqdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-79fpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.881005 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.881066 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.881081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.881114 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.881129 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.889587 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9klkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cdd0c58-7581-4428-b193-2666a9712555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T06:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6cb8c76d04b8ecc5296c008259c091f66f7cf04f973a3cf90fdba2409c16da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T06:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e8f57f1eae73daf6478baad7e94f0d601b6b3090fcefea9aa885d2e9814922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a294b9aa5393598b082562a52bf5fd291ffcddb32e213550f01c388146fb8cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb161ea3e2a0c4fe7ea8d683e9bf48b661d9161e7a54bc7886d570b826b45551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e596badc7048fcde38a70e2a6ea60e899d2f8aeaf70cd6f3078a30df2c1393e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b110c52b69389d0fb520f35fd175244d87e16bd874ba41f2345328e78ce6a4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd1af49c988c50fd4f280580cf9145df793e3d1b6e05cd6da77331fc3daf650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T06:11:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T06:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfjdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T06:11:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9klkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T06:12:31Z is after 2025-08-24T17:21:41Z" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.983620 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.983678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.983690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.983716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.983729 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:31Z","lastTransitionTime":"2025-12-04T06:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:31 crc kubenswrapper[4685]: I1204 06:12:31.985100 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:12:31 crc kubenswrapper[4685]: E1204 06:12:31.985295 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.985258004 +0000 UTC m=+153.173488949 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086450 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086525 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086555 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086583 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086596 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086605 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086650 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.086630013 +0000 UTC m=+153.274860798 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086682 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086691 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086742 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086749 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.086781 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086797 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086805 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086829 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.086805429 +0000 UTC m=+153.275036244 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086836 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086856 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.08684293 +0000 UTC m=+153.275073745 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086857 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.086918 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.086906872 +0000 UTC m=+153.275137687 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.125373 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.125401 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.125559 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.125563 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.125742 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.125750 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.125850 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.126025 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.139750 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.189887 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.189982 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.190008 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.190041 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.190063 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.293684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.294247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.294273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.294304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.294328 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.396850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.396912 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.396930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.396954 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.396973 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.499127 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.499224 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.499245 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.499270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.499287 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.602160 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.602236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.602259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.602288 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.602310 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.614267 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/3.log" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.615227 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/2.log" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.619684 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" exitCode=1 Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.619822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.619931 4685 scope.go:117] "RemoveContainer" containerID="d7ce75633bd88cf4ecc3d408f0899dbcb621aa5e39c262a8328e0e47484d62e7" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.621917 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:12:32 crc kubenswrapper[4685]: E1204 06:12:32.622215 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.711533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.711566 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.711607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.711631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.711646 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.815171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.815233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.815251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.815277 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.815294 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.907600 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrbg8" podStartSLOduration=58.90749742 podStartE2EDuration="58.90749742s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:32.881498283 +0000 UTC m=+90.069729058" watchObservedRunningTime="2025-12-04 06:12:32.90749742 +0000 UTC m=+90.095728235" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.918139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.918197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.918208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.918232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.918247 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:32Z","lastTransitionTime":"2025-12-04T06:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.921617 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=0.921599566 podStartE2EDuration="921.599566ms" podCreationTimestamp="2025-12-04 06:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:32.921337187 +0000 UTC m=+90.109567962" watchObservedRunningTime="2025-12-04 06:12:32.921599566 +0000 UTC m=+90.109830341" Dec 04 06:12:32 crc kubenswrapper[4685]: I1204 06:12:32.966862 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=64.966832062 podStartE2EDuration="1m4.966832062s" podCreationTimestamp="2025-12-04 06:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:32.948058188 +0000 UTC m=+90.136288963" watchObservedRunningTime="2025-12-04 06:12:32.966832062 +0000 UTC m=+90.155062877" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.014204 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podStartSLOduration=60.014182199 podStartE2EDuration="1m0.014182199s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:32.994750973 +0000 UTC m=+90.182981748" watchObservedRunningTime="2025-12-04 06:12:33.014182199 +0000 UTC m=+90.202412984" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.020691 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.020737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.020748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.020764 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.020777 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.031575 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-79fpc" podStartSLOduration=60.031549564 podStartE2EDuration="1m0.031549564s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:33.031535163 +0000 UTC m=+90.219765948" watchObservedRunningTime="2025-12-04 06:12:33.031549564 +0000 UTC m=+90.219780349" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.048216 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9klkp" podStartSLOduration=60.048194475 podStartE2EDuration="1m0.048194475s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:33.047025346 +0000 UTC m=+90.235256121" watchObservedRunningTime="2025-12-04 06:12:33.048194475 +0000 UTC m=+90.236425250" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.081847 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.081821779 podStartE2EDuration="41.081821779s" podCreationTimestamp="2025-12-04 06:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:33.08063466 +0000 UTC m=+90.268865435" watchObservedRunningTime="2025-12-04 06:12:33.081821779 +0000 UTC m=+90.270052584" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.106179 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-dpd9d" podStartSLOduration=60.106161761 podStartE2EDuration="1m0.106161761s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:33.105962914 +0000 UTC m=+90.294193689" watchObservedRunningTime="2025-12-04 06:12:33.106161761 +0000 UTC m=+90.294392536" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.123402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.123451 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.123467 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.123483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.123494 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.141989 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-47k4t" podStartSLOduration=60.141970568 podStartE2EDuration="1m0.141970568s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:33.14113286 +0000 UTC m=+90.329363645" watchObservedRunningTime="2025-12-04 06:12:33.141970568 +0000 UTC m=+90.330201363" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.154781 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=61.154763379 podStartE2EDuration="1m1.154763379s" podCreationTimestamp="2025-12-04 06:11:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:33.154212392 +0000 UTC m=+90.342443177" watchObservedRunningTime="2025-12-04 06:12:33.154763379 +0000 UTC m=+90.342994164" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.226069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.226454 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.226558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.226582 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.226595 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.329184 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.329224 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.329236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.329254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.329266 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.432027 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.432478 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.432711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.432961 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.433182 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.537313 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.537439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.537466 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.537502 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.537526 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.626967 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/3.log" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.632832 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:12:33 crc kubenswrapper[4685]: E1204 06:12:33.633091 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.642526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.642579 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.642598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.642623 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.642640 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.747509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.747581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.747598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.748001 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.748057 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.851220 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.851287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.851304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.851772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.851829 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.954394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.954714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.954806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.955034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:33 crc kubenswrapper[4685]: I1204 06:12:33.955146 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:33Z","lastTransitionTime":"2025-12-04T06:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.058878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.058935 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.058954 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.058979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.058997 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.125481 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.125522 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.125964 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:34 crc kubenswrapper[4685]: E1204 06:12:34.126194 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.126362 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:34 crc kubenswrapper[4685]: E1204 06:12:34.126513 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:34 crc kubenswrapper[4685]: E1204 06:12:34.126372 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:34 crc kubenswrapper[4685]: E1204 06:12:34.126830 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.162559 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.162935 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.163118 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.163290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.163504 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.266843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.266922 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.266948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.266976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.266996 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.369904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.369949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.369960 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.369978 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.369989 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.473159 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.473217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.473242 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.473272 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.473296 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.576029 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.576092 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.576149 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.576178 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.576198 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.679004 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.679074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.679091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.679116 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.679135 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.782873 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.782941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.782958 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.782985 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.783002 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.886332 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.886394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.886446 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.886478 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.886533 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.989527 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.989588 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.989607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.989631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:34 crc kubenswrapper[4685]: I1204 06:12:34.989649 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:34Z","lastTransitionTime":"2025-12-04T06:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.092586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.092633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.092646 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.092662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.092675 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.195074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.195113 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.195124 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.195139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.195150 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.298404 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.298551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.298588 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.298617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.298638 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.402610 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.402670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.402687 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.402711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.402730 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.505771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.505878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.505958 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.506880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.506965 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.609201 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.609645 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.609863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.610061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.610280 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.713341 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.713529 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.713555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.713584 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.713604 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.816690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.816732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.816752 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.816781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.816805 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.919940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.920001 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.920019 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.920045 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:35 crc kubenswrapper[4685]: I1204 06:12:35.920063 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:35Z","lastTransitionTime":"2025-12-04T06:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.023692 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.024098 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.024233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.024354 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.024514 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.124859 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.124947 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.124981 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:36 crc kubenswrapper[4685]: E1204 06:12:36.125359 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:36 crc kubenswrapper[4685]: E1204 06:12:36.125887 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:36 crc kubenswrapper[4685]: E1204 06:12:36.126111 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.127184 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.127225 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.127248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.127277 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.127301 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.127779 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:36 crc kubenswrapper[4685]: E1204 06:12:36.128036 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.230302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.230358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.230375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.230401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.230448 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.333661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.334922 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.334967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.334996 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.335015 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.441465 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.441535 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.441558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.441592 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.441614 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.545155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.545651 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.545836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.545996 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.546135 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.648309 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.648345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.648354 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.648370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.648380 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.751456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.751526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.751539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.751563 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.751577 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.854946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.855319 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.855509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.855802 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.855950 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.958481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.958903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.959035 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.959180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:36 crc kubenswrapper[4685]: I1204 06:12:36.959324 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:36Z","lastTransitionTime":"2025-12-04T06:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.064190 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.064282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.064307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.064340 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.064365 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.168099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.168176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.168199 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.168231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.168254 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.271585 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.271617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.271628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.271644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.271710 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.374352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.374464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.374489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.374519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.374536 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.477376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.477467 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.477479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.477512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.477523 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.579917 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.579989 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.580007 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.580037 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.580055 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.682617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.682688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.682710 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.682737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.682758 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.786142 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.786192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.786210 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.786234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.786252 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.890234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.890314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.890336 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.890370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.890390 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.993121 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.993179 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.993195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.993217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:37 crc kubenswrapper[4685]: I1204 06:12:37.993234 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:37Z","lastTransitionTime":"2025-12-04T06:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.096981 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.097052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.097070 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.097095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.097114 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.125559 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.125628 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.125633 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:38 crc kubenswrapper[4685]: E1204 06:12:38.125791 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.125893 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:38 crc kubenswrapper[4685]: E1204 06:12:38.126060 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:38 crc kubenswrapper[4685]: E1204 06:12:38.126393 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:38 crc kubenswrapper[4685]: E1204 06:12:38.126600 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.200656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.200750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.200768 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.201232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.201629 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.304721 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.304817 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.304836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.304859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.304881 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.408125 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.408207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.408228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.408253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.408271 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.511959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.512018 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.512038 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.512061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.512077 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.615063 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.615138 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.615154 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.615178 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.615195 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.719784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.719899 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.719916 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.719940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.719956 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.822486 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.822559 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.822577 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.822604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.822624 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.925365 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.925430 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.925485 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.925510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:38 crc kubenswrapper[4685]: I1204 06:12:38.925529 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:38Z","lastTransitionTime":"2025-12-04T06:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.028179 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.028266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.028300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.028329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.028348 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.131522 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.131580 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.131596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.131619 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.131640 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.235441 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.235509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.235526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.235555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.235576 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.338839 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.338906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.338925 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.338950 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.338967 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.442843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.442919 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.442941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.442969 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.442992 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.546250 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.546346 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.546400 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.546463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.546489 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.605357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.605471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.605491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.605516 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.605534 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T06:12:39Z","lastTransitionTime":"2025-12-04T06:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.660988 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8"] Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.661324 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.663001 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.663234 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.664024 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.664145 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.788629 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.788678 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.788877 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.788981 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.789044 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890008 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890096 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890146 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890182 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890216 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890729 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.890843 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.891784 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.903093 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.916871 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5b5x8\" (UID: \"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:39 crc kubenswrapper[4685]: I1204 06:12:39.978049 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.125821 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:40 crc kubenswrapper[4685]: E1204 06:12:40.126018 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.126307 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:40 crc kubenswrapper[4685]: E1204 06:12:40.126424 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.126586 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.126636 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:40 crc kubenswrapper[4685]: E1204 06:12:40.128099 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:40 crc kubenswrapper[4685]: E1204 06:12:40.127979 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.148928 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.655544 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" event={"ID":"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed","Type":"ContainerStarted","Data":"0edbf6498d8d943e6c0d33d4a0cbe3d5a8e15f1a26ea25a917f5f047b26f542a"} Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.655603 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" event={"ID":"93b3f32b-7c1e-4d96-8de4-f7ffa769e1ed","Type":"ContainerStarted","Data":"3a89cd06fd10860ac7218b60ef81445fe4391aaeb9c697d10874b05c451111cb"} Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.711956 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5b5x8" podStartSLOduration=67.711929614 podStartE2EDuration="1m7.711929614s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:40.676418396 +0000 UTC m=+97.864649181" watchObservedRunningTime="2025-12-04 06:12:40.711929614 +0000 UTC m=+97.900160399" Dec 04 06:12:40 crc kubenswrapper[4685]: I1204 06:12:40.712587 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=0.712579706 podStartE2EDuration="712.579706ms" podCreationTimestamp="2025-12-04 06:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:12:40.711149027 +0000 UTC m=+97.899379802" watchObservedRunningTime="2025-12-04 06:12:40.712579706 +0000 UTC m=+97.900810511" Dec 04 06:12:42 crc kubenswrapper[4685]: I1204 06:12:42.124764 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:42 crc kubenswrapper[4685]: I1204 06:12:42.124880 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:42 crc kubenswrapper[4685]: I1204 06:12:42.124767 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:42 crc kubenswrapper[4685]: E1204 06:12:42.124954 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:42 crc kubenswrapper[4685]: E1204 06:12:42.125129 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:42 crc kubenswrapper[4685]: E1204 06:12:42.125331 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:42 crc kubenswrapper[4685]: I1204 06:12:42.125505 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:42 crc kubenswrapper[4685]: E1204 06:12:42.125649 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:44 crc kubenswrapper[4685]: I1204 06:12:44.124610 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:44 crc kubenswrapper[4685]: I1204 06:12:44.124718 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:44 crc kubenswrapper[4685]: E1204 06:12:44.124801 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:44 crc kubenswrapper[4685]: I1204 06:12:44.124848 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:44 crc kubenswrapper[4685]: I1204 06:12:44.124906 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:44 crc kubenswrapper[4685]: E1204 06:12:44.125030 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:44 crc kubenswrapper[4685]: E1204 06:12:44.125179 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:44 crc kubenswrapper[4685]: E1204 06:12:44.125262 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:46 crc kubenswrapper[4685]: I1204 06:12:46.125458 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:46 crc kubenswrapper[4685]: I1204 06:12:46.125537 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:46 crc kubenswrapper[4685]: I1204 06:12:46.125471 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:46 crc kubenswrapper[4685]: E1204 06:12:46.125700 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:46 crc kubenswrapper[4685]: I1204 06:12:46.125731 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:46 crc kubenswrapper[4685]: E1204 06:12:46.125878 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:46 crc kubenswrapper[4685]: E1204 06:12:46.125993 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:46 crc kubenswrapper[4685]: E1204 06:12:46.126088 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:48 crc kubenswrapper[4685]: I1204 06:12:48.124682 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:48 crc kubenswrapper[4685]: I1204 06:12:48.124792 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:48 crc kubenswrapper[4685]: E1204 06:12:48.124878 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:48 crc kubenswrapper[4685]: I1204 06:12:48.124686 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:48 crc kubenswrapper[4685]: I1204 06:12:48.124720 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:48 crc kubenswrapper[4685]: E1204 06:12:48.125052 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:48 crc kubenswrapper[4685]: E1204 06:12:48.125192 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:48 crc kubenswrapper[4685]: E1204 06:12:48.125396 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:49 crc kubenswrapper[4685]: I1204 06:12:49.126701 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:12:49 crc kubenswrapper[4685]: E1204 06:12:49.127712 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:12:50 crc kubenswrapper[4685]: I1204 06:12:50.125688 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:50 crc kubenswrapper[4685]: I1204 06:12:50.125867 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:50 crc kubenswrapper[4685]: I1204 06:12:50.126068 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:50 crc kubenswrapper[4685]: E1204 06:12:50.126063 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:50 crc kubenswrapper[4685]: I1204 06:12:50.126120 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:50 crc kubenswrapper[4685]: E1204 06:12:50.126327 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:50 crc kubenswrapper[4685]: E1204 06:12:50.126569 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:50 crc kubenswrapper[4685]: E1204 06:12:50.126640 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:52 crc kubenswrapper[4685]: I1204 06:12:52.125774 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:52 crc kubenswrapper[4685]: E1204 06:12:52.125972 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:52 crc kubenswrapper[4685]: I1204 06:12:52.126350 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:52 crc kubenswrapper[4685]: E1204 06:12:52.126525 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:52 crc kubenswrapper[4685]: I1204 06:12:52.126677 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:52 crc kubenswrapper[4685]: E1204 06:12:52.126847 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:52 crc kubenswrapper[4685]: I1204 06:12:52.126869 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:52 crc kubenswrapper[4685]: E1204 06:12:52.126997 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:52 crc kubenswrapper[4685]: I1204 06:12:52.228734 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:52 crc kubenswrapper[4685]: E1204 06:12:52.228921 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:12:52 crc kubenswrapper[4685]: E1204 06:12:52.229045 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs podName:06bbc9d6-e56b-4c35-bdd7-659ab3d3a870 nodeName:}" failed. No retries permitted until 2025-12-04 06:13:56.229011201 +0000 UTC m=+173.417242016 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs") pod "network-metrics-daemon-55csn" (UID: "06bbc9d6-e56b-4c35-bdd7-659ab3d3a870") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 06:12:54 crc kubenswrapper[4685]: I1204 06:12:54.124681 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:54 crc kubenswrapper[4685]: E1204 06:12:54.125214 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:54 crc kubenswrapper[4685]: I1204 06:12:54.124845 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:54 crc kubenswrapper[4685]: E1204 06:12:54.125343 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:54 crc kubenswrapper[4685]: I1204 06:12:54.124831 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:54 crc kubenswrapper[4685]: I1204 06:12:54.124887 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:54 crc kubenswrapper[4685]: E1204 06:12:54.125513 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:54 crc kubenswrapper[4685]: E1204 06:12:54.125686 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:56 crc kubenswrapper[4685]: I1204 06:12:56.127888 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:56 crc kubenswrapper[4685]: I1204 06:12:56.127933 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:56 crc kubenswrapper[4685]: I1204 06:12:56.127904 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:56 crc kubenswrapper[4685]: I1204 06:12:56.127934 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:56 crc kubenswrapper[4685]: E1204 06:12:56.128210 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:56 crc kubenswrapper[4685]: E1204 06:12:56.128360 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:12:56 crc kubenswrapper[4685]: E1204 06:12:56.128645 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:56 crc kubenswrapper[4685]: E1204 06:12:56.128974 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:58 crc kubenswrapper[4685]: I1204 06:12:58.125342 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:12:58 crc kubenswrapper[4685]: I1204 06:12:58.125390 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:12:58 crc kubenswrapper[4685]: I1204 06:12:58.125470 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:12:58 crc kubenswrapper[4685]: I1204 06:12:58.125390 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:12:58 crc kubenswrapper[4685]: E1204 06:12:58.125644 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:12:58 crc kubenswrapper[4685]: E1204 06:12:58.125808 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:12:58 crc kubenswrapper[4685]: E1204 06:12:58.125893 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:12:58 crc kubenswrapper[4685]: E1204 06:12:58.126019 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:00 crc kubenswrapper[4685]: I1204 06:13:00.125066 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:00 crc kubenswrapper[4685]: I1204 06:13:00.125223 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:00 crc kubenswrapper[4685]: I1204 06:13:00.125259 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:00 crc kubenswrapper[4685]: I1204 06:13:00.125295 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:00 crc kubenswrapper[4685]: E1204 06:13:00.125629 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:00 crc kubenswrapper[4685]: E1204 06:13:00.125776 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:00 crc kubenswrapper[4685]: E1204 06:13:00.125920 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:00 crc kubenswrapper[4685]: E1204 06:13:00.126037 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:02 crc kubenswrapper[4685]: I1204 06:13:02.124634 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:02 crc kubenswrapper[4685]: I1204 06:13:02.124707 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:02 crc kubenswrapper[4685]: E1204 06:13:02.124871 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:02 crc kubenswrapper[4685]: I1204 06:13:02.124898 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:02 crc kubenswrapper[4685]: E1204 06:13:02.125039 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:02 crc kubenswrapper[4685]: E1204 06:13:02.125343 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:02 crc kubenswrapper[4685]: I1204 06:13:02.125489 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:02 crc kubenswrapper[4685]: E1204 06:13:02.125646 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:03 crc kubenswrapper[4685]: E1204 06:13:03.079826 4685 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 06:13:03 crc kubenswrapper[4685]: I1204 06:13:03.128202 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:13:03 crc kubenswrapper[4685]: E1204 06:13:03.128507 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n4rjr_openshift-ovn-kubernetes(f4c7de36-abbd-448e-bf66-eecafc38068c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" Dec 04 06:13:03 crc kubenswrapper[4685]: E1204 06:13:03.192723 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 06:13:04 crc kubenswrapper[4685]: I1204 06:13:04.125472 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:04 crc kubenswrapper[4685]: I1204 06:13:04.125474 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:04 crc kubenswrapper[4685]: I1204 06:13:04.125472 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:04 crc kubenswrapper[4685]: I1204 06:13:04.125640 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:04 crc kubenswrapper[4685]: E1204 06:13:04.125890 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:04 crc kubenswrapper[4685]: E1204 06:13:04.126087 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:04 crc kubenswrapper[4685]: E1204 06:13:04.126211 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:04 crc kubenswrapper[4685]: E1204 06:13:04.126374 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:06 crc kubenswrapper[4685]: I1204 06:13:06.125548 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:06 crc kubenswrapper[4685]: I1204 06:13:06.125598 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:06 crc kubenswrapper[4685]: I1204 06:13:06.125631 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:06 crc kubenswrapper[4685]: I1204 06:13:06.125845 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:06 crc kubenswrapper[4685]: E1204 06:13:06.125994 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:06 crc kubenswrapper[4685]: E1204 06:13:06.126132 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:06 crc kubenswrapper[4685]: E1204 06:13:06.126259 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:06 crc kubenswrapper[4685]: E1204 06:13:06.126371 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:07 crc kubenswrapper[4685]: I1204 06:13:07.749988 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/1.log" Dec 04 06:13:07 crc kubenswrapper[4685]: I1204 06:13:07.751016 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/0.log" Dec 04 06:13:07 crc kubenswrapper[4685]: I1204 06:13:07.751088 4685 generic.go:334] "Generic (PLEG): container finished" podID="736d38b0-c0ce-4db5-bee6-72ee0dc23b40" containerID="b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643" exitCode=1 Dec 04 06:13:07 crc kubenswrapper[4685]: I1204 06:13:07.751138 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerDied","Data":"b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643"} Dec 04 06:13:07 crc kubenswrapper[4685]: I1204 06:13:07.751206 4685 scope.go:117] "RemoveContainer" containerID="b219c5eb80b6d2af4737f6807196747d6e3e88f9355f5875ff0dee271c350de3" Dec 04 06:13:07 crc kubenswrapper[4685]: I1204 06:13:07.751711 4685 scope.go:117] "RemoveContainer" containerID="b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643" Dec 04 06:13:07 crc kubenswrapper[4685]: E1204 06:13:07.752006 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-dpd9d_openshift-multus(736d38b0-c0ce-4db5-bee6-72ee0dc23b40)\"" pod="openshift-multus/multus-dpd9d" podUID="736d38b0-c0ce-4db5-bee6-72ee0dc23b40" Dec 04 06:13:08 crc kubenswrapper[4685]: I1204 06:13:08.124808 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:08 crc kubenswrapper[4685]: I1204 06:13:08.124892 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:08 crc kubenswrapper[4685]: I1204 06:13:08.124920 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:08 crc kubenswrapper[4685]: E1204 06:13:08.124961 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:08 crc kubenswrapper[4685]: I1204 06:13:08.124983 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:08 crc kubenswrapper[4685]: E1204 06:13:08.125190 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:08 crc kubenswrapper[4685]: E1204 06:13:08.125324 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:08 crc kubenswrapper[4685]: E1204 06:13:08.125494 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:08 crc kubenswrapper[4685]: E1204 06:13:08.194291 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 06:13:08 crc kubenswrapper[4685]: I1204 06:13:08.755910 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/1.log" Dec 04 06:13:10 crc kubenswrapper[4685]: I1204 06:13:10.125587 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:10 crc kubenswrapper[4685]: I1204 06:13:10.125641 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:10 crc kubenswrapper[4685]: I1204 06:13:10.125687 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:10 crc kubenswrapper[4685]: E1204 06:13:10.125722 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:10 crc kubenswrapper[4685]: I1204 06:13:10.125745 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:10 crc kubenswrapper[4685]: E1204 06:13:10.125869 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:10 crc kubenswrapper[4685]: E1204 06:13:10.125958 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:10 crc kubenswrapper[4685]: E1204 06:13:10.126012 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:12 crc kubenswrapper[4685]: I1204 06:13:12.124578 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:12 crc kubenswrapper[4685]: I1204 06:13:12.124658 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:12 crc kubenswrapper[4685]: E1204 06:13:12.124725 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:12 crc kubenswrapper[4685]: I1204 06:13:12.124815 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:12 crc kubenswrapper[4685]: I1204 06:13:12.124598 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:12 crc kubenswrapper[4685]: E1204 06:13:12.124867 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:12 crc kubenswrapper[4685]: E1204 06:13:12.125023 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:12 crc kubenswrapper[4685]: E1204 06:13:12.125159 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:13 crc kubenswrapper[4685]: E1204 06:13:13.194838 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.125514 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.125515 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.125584 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.125762 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:14 crc kubenswrapper[4685]: E1204 06:13:14.125945 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:14 crc kubenswrapper[4685]: E1204 06:13:14.126036 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:14 crc kubenswrapper[4685]: E1204 06:13:14.126170 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.126230 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:13:14 crc kubenswrapper[4685]: E1204 06:13:14.126295 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.775855 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/3.log" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.778270 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerStarted","Data":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.778661 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.803986 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podStartSLOduration=101.803962366 podStartE2EDuration="1m41.803962366s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:14.803825821 +0000 UTC m=+131.992056596" watchObservedRunningTime="2025-12-04 06:13:14.803962366 +0000 UTC m=+131.992193161" Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.956742 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-55csn"] Dec 04 06:13:14 crc kubenswrapper[4685]: I1204 06:13:14.956893 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:14 crc kubenswrapper[4685]: E1204 06:13:14.957047 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:16 crc kubenswrapper[4685]: I1204 06:13:16.125510 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:16 crc kubenswrapper[4685]: I1204 06:13:16.125558 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:16 crc kubenswrapper[4685]: I1204 06:13:16.125648 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:16 crc kubenswrapper[4685]: E1204 06:13:16.126256 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:16 crc kubenswrapper[4685]: E1204 06:13:16.126490 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:16 crc kubenswrapper[4685]: E1204 06:13:16.126047 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:17 crc kubenswrapper[4685]: I1204 06:13:17.125696 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:17 crc kubenswrapper[4685]: E1204 06:13:17.125830 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:18 crc kubenswrapper[4685]: I1204 06:13:18.124669 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:18 crc kubenswrapper[4685]: I1204 06:13:18.124759 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:18 crc kubenswrapper[4685]: E1204 06:13:18.124806 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:18 crc kubenswrapper[4685]: I1204 06:13:18.124848 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:18 crc kubenswrapper[4685]: E1204 06:13:18.125002 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:18 crc kubenswrapper[4685]: E1204 06:13:18.125092 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:18 crc kubenswrapper[4685]: E1204 06:13:18.196743 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 06:13:19 crc kubenswrapper[4685]: I1204 06:13:19.125443 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:19 crc kubenswrapper[4685]: E1204 06:13:19.125557 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:20 crc kubenswrapper[4685]: I1204 06:13:20.124802 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:20 crc kubenswrapper[4685]: I1204 06:13:20.124839 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:20 crc kubenswrapper[4685]: E1204 06:13:20.124935 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:20 crc kubenswrapper[4685]: I1204 06:13:20.124955 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:20 crc kubenswrapper[4685]: E1204 06:13:20.125026 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:20 crc kubenswrapper[4685]: E1204 06:13:20.125103 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:21 crc kubenswrapper[4685]: I1204 06:13:21.125535 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:21 crc kubenswrapper[4685]: E1204 06:13:21.125660 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:22 crc kubenswrapper[4685]: I1204 06:13:22.124793 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:22 crc kubenswrapper[4685]: I1204 06:13:22.124834 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:22 crc kubenswrapper[4685]: I1204 06:13:22.124881 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:22 crc kubenswrapper[4685]: E1204 06:13:22.124965 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:22 crc kubenswrapper[4685]: E1204 06:13:22.125127 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:22 crc kubenswrapper[4685]: E1204 06:13:22.125222 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:22 crc kubenswrapper[4685]: I1204 06:13:22.125499 4685 scope.go:117] "RemoveContainer" containerID="b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643" Dec 04 06:13:22 crc kubenswrapper[4685]: I1204 06:13:22.808031 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/1.log" Dec 04 06:13:22 crc kubenswrapper[4685]: I1204 06:13:22.808102 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerStarted","Data":"23cbb5fd93c677593365a3111be3cf45e952008b1d7520ce81f4f2b7c902f1eb"} Dec 04 06:13:23 crc kubenswrapper[4685]: I1204 06:13:23.125282 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:23 crc kubenswrapper[4685]: E1204 06:13:23.126678 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:23 crc kubenswrapper[4685]: E1204 06:13:23.197384 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 06:13:24 crc kubenswrapper[4685]: I1204 06:13:24.125255 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:24 crc kubenswrapper[4685]: E1204 06:13:24.125393 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:24 crc kubenswrapper[4685]: I1204 06:13:24.125550 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:24 crc kubenswrapper[4685]: I1204 06:13:24.125618 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:24 crc kubenswrapper[4685]: E1204 06:13:24.125687 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:24 crc kubenswrapper[4685]: E1204 06:13:24.125861 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:25 crc kubenswrapper[4685]: I1204 06:13:25.125362 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:25 crc kubenswrapper[4685]: E1204 06:13:25.125791 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:26 crc kubenswrapper[4685]: I1204 06:13:26.124550 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:26 crc kubenswrapper[4685]: I1204 06:13:26.124617 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:26 crc kubenswrapper[4685]: E1204 06:13:26.124690 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:26 crc kubenswrapper[4685]: I1204 06:13:26.124835 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:26 crc kubenswrapper[4685]: E1204 06:13:26.125044 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:26 crc kubenswrapper[4685]: E1204 06:13:26.125309 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:27 crc kubenswrapper[4685]: I1204 06:13:27.127300 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:27 crc kubenswrapper[4685]: E1204 06:13:27.127422 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55csn" podUID="06bbc9d6-e56b-4c35-bdd7-659ab3d3a870" Dec 04 06:13:28 crc kubenswrapper[4685]: I1204 06:13:28.125346 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:28 crc kubenswrapper[4685]: I1204 06:13:28.125359 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:28 crc kubenswrapper[4685]: E1204 06:13:28.125572 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 06:13:28 crc kubenswrapper[4685]: E1204 06:13:28.125697 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 06:13:28 crc kubenswrapper[4685]: I1204 06:13:28.125366 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:28 crc kubenswrapper[4685]: E1204 06:13:28.125815 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 06:13:29 crc kubenswrapper[4685]: I1204 06:13:29.125449 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:29 crc kubenswrapper[4685]: I1204 06:13:29.128038 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 06:13:29 crc kubenswrapper[4685]: I1204 06:13:29.128470 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.125382 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.125816 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.125842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.127096 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.128214 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.128912 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.129375 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.939044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.983182 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fz7dx"] Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.983764 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.985265 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b9hw7"] Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.985593 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8"] Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.985900 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.986067 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qkqs"] Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.986208 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.986561 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.986723 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.987544 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5"] Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.988002 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.992112 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.993179 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.993495 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cplqg"] Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.993881 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.994224 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.994786 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 06:13:30 crc kubenswrapper[4685]: I1204 06:13:30.995095 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.043963 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-24w48"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.044551 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.045022 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.045364 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.045626 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.045886 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.046512 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.046753 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.047504 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rfvkg"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.047978 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.055210 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.055309 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.055625 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.058892 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.058914 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.059277 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.059315 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.059441 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.059509 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.059626 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.060668 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.068378 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.076593 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.077569 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.077719 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.077935 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.078121 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.078542 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.080748 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.080926 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.081059 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.081292 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.082176 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-s6kbn"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.085497 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vkf95"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.087905 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.088270 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.094269 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.087947 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.083563 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.083615 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.087998 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.088051 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.088186 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.082224 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.096016 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.096268 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.119810 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.120267 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.120682 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.120843 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121147 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121224 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121293 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121359 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121459 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121605 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.121680 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.122090 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.122201 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.122550 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6zcm5"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.122871 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.123216 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.123562 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.123596 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.123692 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.123828 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.138613 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.138799 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.138898 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.139060 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.139147 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.141862 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7l82d"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.142254 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.142465 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.142747 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.143274 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.158507 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.159110 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jkj6h"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.160574 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b9hw7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.160689 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.158509 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-config\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161434 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89e148da-1707-4caa-9e28-7976df0303a5-audit-dir\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161485 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161522 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25743472-93a4-4d4e-bf54-1d5cef4025a2-config\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161592 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00e041f6-1d07-4dfe-a844-ed090a258aec-serving-cert\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161618 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv8m9\" (UniqueName: \"kubernetes.io/projected/6dd8d948-2f17-4ef8-8b31-572f08420bb2-kube-api-access-qv8m9\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161652 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6dd8d948-2f17-4ef8-8b31-572f08420bb2-serving-cert\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161676 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-policies\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161754 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161775 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25743472-93a4-4d4e-bf54-1d5cef4025a2-images\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161809 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bf8ca363-9882-4e15-bfe7-5b59686e88ea-node-pullsecrets\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161853 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161914 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc2k7\" (UniqueName: \"kubernetes.io/projected/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-kube-api-access-rc2k7\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161932 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-service-ca-bundle\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.161996 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-dir\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162022 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162132 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-config\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162164 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6dd8d948-2f17-4ef8-8b31-572f08420bb2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162179 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-etcd-serving-ca\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162198 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162278 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-audit-policies\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162295 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crx5m\" (UniqueName: \"kubernetes.io/projected/bf8ca363-9882-4e15-bfe7-5b59686e88ea-kube-api-access-crx5m\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162335 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-serving-cert\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162352 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms5nw\" (UniqueName: \"kubernetes.io/projected/25743472-93a4-4d4e-bf54-1d5cef4025a2-kube-api-access-ms5nw\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162423 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162441 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9cl2\" (UniqueName: \"kubernetes.io/projected/b50783dc-85ca-4dd8-9ad7-08224cf4934f-kube-api-access-k9cl2\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162506 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-encryption-config\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162567 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-etcd-client\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162595 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162638 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162657 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-audit\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162706 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162776 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162801 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-encryption-config\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162849 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162899 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162918 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.162985 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4tx8\" (UniqueName: \"kubernetes.io/projected/89e148da-1707-4caa-9e28-7976df0303a5-kube-api-access-d4tx8\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163005 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqdqm\" (UniqueName: \"kubernetes.io/projected/00e041f6-1d07-4dfe-a844-ed090a258aec-kube-api-access-xqdqm\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163109 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163147 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163166 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b50783dc-85ca-4dd8-9ad7-08224cf4934f-serving-cert\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163188 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/25743472-93a4-4d4e-bf54-1d5cef4025a2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163204 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163223 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163350 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-etcd-client\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-serving-cert\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163445 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163543 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-config\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163570 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf8ca363-9882-4e15-bfe7-5b59686e88ea-audit-dir\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163571 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s6zxb"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163599 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-image-import-ca\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163620 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-client-ca\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.163659 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s96jb\" (UniqueName: \"kubernetes.io/projected/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-kube-api-access-s96jb\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.176836 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.177254 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.177433 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.178921 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.179798 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.182716 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.183102 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.183247 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.183258 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.183684 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.183839 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.184064 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.184147 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.184283 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.185035 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.185097 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.185794 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.185897 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.185946 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186037 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186154 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186265 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186364 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186473 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186597 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186620 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186676 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.187016 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.186729 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.187759 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.190273 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.190379 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.191882 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.192144 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-qxx2n"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.192458 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.192699 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.192789 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.193066 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.193095 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.193132 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.194942 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.195120 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.195202 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fz7dx"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.195233 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.195331 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.195669 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.195805 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.196280 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.197118 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.197737 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.198005 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.201907 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.202248 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.212694 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.212733 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.212780 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.212733 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.212953 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.213011 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.213098 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.213275 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.214700 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.215752 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qkqs"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.215946 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.218302 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.224243 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-24w48"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.232700 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vkf95"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.232760 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cplqg"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.232778 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8ndqp"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.226920 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.233492 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.233625 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.233859 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.233928 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.234260 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.235459 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.236900 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.237288 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.238128 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.239302 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.239735 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6zcm5"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.240660 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.241394 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.241519 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.242565 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.243053 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.243215 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.243894 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q4sb7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.244319 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.244434 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.244833 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.245902 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.246624 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.246686 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.247161 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crn6w"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.247864 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.248085 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.248719 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.249054 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rfvkg"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.250275 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s6zxb"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.251541 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.252911 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.253534 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-g9csq"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.254116 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.254269 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.255230 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.256622 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.258184 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.259146 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s6kbn"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.260119 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7l82d"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.261098 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.262072 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.263028 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jkj6h"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264104 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8ndqp"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264451 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-encryption-config\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264481 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264503 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9cl2\" (UniqueName: \"kubernetes.io/projected/b50783dc-85ca-4dd8-9ad7-08224cf4934f-kube-api-access-k9cl2\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264523 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-etcd-client\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264547 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264564 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264578 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264595 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-encryption-config\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-audit\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264625 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264645 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1880937-0dcd-437c-8993-021954821db3-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264664 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1880937-0dcd-437c-8993-021954821db3-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264680 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4tx8\" (UniqueName: \"kubernetes.io/projected/89e148da-1707-4caa-9e28-7976df0303a5-kube-api-access-d4tx8\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264697 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264712 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264730 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264751 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqdqm\" (UniqueName: \"kubernetes.io/projected/00e041f6-1d07-4dfe-a844-ed090a258aec-kube-api-access-xqdqm\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264769 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264786 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqjdr\" (UniqueName: \"kubernetes.io/projected/f53d1c26-cef7-4963-aebc-bb7d8489f070-kube-api-access-pqjdr\") pod \"dns-operator-744455d44c-7l82d\" (UID: \"f53d1c26-cef7-4963-aebc-bb7d8489f070\") " pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264802 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264819 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b50783dc-85ca-4dd8-9ad7-08224cf4934f-serving-cert\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264835 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-etcd-client\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264855 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/25743472-93a4-4d4e-bf54-1d5cef4025a2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264872 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264888 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264905 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-serving-cert\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264921 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-config\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264939 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264955 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf8ca363-9882-4e15-bfe7-5b59686e88ea-audit-dir\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264983 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-image-import-ca\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.264998 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-client-ca\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265015 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s96jb\" (UniqueName: \"kubernetes.io/projected/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-kube-api-access-s96jb\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265037 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-config\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265053 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265071 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89e148da-1707-4caa-9e28-7976df0303a5-audit-dir\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265086 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25743472-93a4-4d4e-bf54-1d5cef4025a2-config\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00e041f6-1d07-4dfe-a844-ed090a258aec-serving-cert\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265120 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv8m9\" (UniqueName: \"kubernetes.io/projected/6dd8d948-2f17-4ef8-8b31-572f08420bb2-kube-api-access-qv8m9\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265136 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1880937-0dcd-437c-8993-021954821db3-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265153 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6dd8d948-2f17-4ef8-8b31-572f08420bb2-serving-cert\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-policies\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265187 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265202 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25743472-93a4-4d4e-bf54-1d5cef4025a2-images\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265217 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bf8ca363-9882-4e15-bfe7-5b59686e88ea-node-pullsecrets\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265232 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f53d1c26-cef7-4963-aebc-bb7d8489f070-metrics-tls\") pod \"dns-operator-744455d44c-7l82d\" (UID: \"f53d1c26-cef7-4963-aebc-bb7d8489f070\") " pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265248 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc2k7\" (UniqueName: \"kubernetes.io/projected/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-kube-api-access-rc2k7\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265264 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-service-ca-bundle\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265296 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265315 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-dir\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265329 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265346 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-config\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265361 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6dd8d948-2f17-4ef8-8b31-572f08420bb2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265377 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-etcd-serving-ca\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265393 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265423 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-audit-policies\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265443 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms5nw\" (UniqueName: \"kubernetes.io/projected/25743472-93a4-4d4e-bf54-1d5cef4025a2-kube-api-access-ms5nw\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265445 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jjm5p"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265463 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crx5m\" (UniqueName: \"kubernetes.io/projected/bf8ca363-9882-4e15-bfe7-5b59686e88ea-kube-api-access-crx5m\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265487 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-serving-cert\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265540 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265682 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.266282 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.266382 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bf8ca363-9882-4e15-bfe7-5b59686e88ea-audit-dir\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.266571 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.266656 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.265448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.266999 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.267197 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-image-import-ca\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.268311 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-client-ca\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269211 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269255 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269284 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-config\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269299 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bf8ca363-9882-4e15-bfe7-5b59686e88ea-node-pullsecrets\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269702 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89e148da-1707-4caa-9e28-7976df0303a5-audit-dir\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269948 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-service-ca-bundle\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269978 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6dd8d948-2f17-4ef8-8b31-572f08420bb2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.269988 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.270149 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.270294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-audit\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.270476 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bf8ca363-9882-4e15-bfe7-5b59686e88ea-etcd-serving-ca\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.270890 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.270916 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25743472-93a4-4d4e-bf54-1d5cef4025a2-images\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.270945 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.273730 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-config\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.273802 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89e148da-1707-4caa-9e28-7976df0303a5-audit-policies\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.273954 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-encryption-config\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274087 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274276 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274274 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00e041f6-1d07-4dfe-a844-ed090a258aec-serving-cert\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274391 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274859 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/25743472-93a4-4d4e-bf54-1d5cef4025a2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274896 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-serving-cert\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.274931 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.275151 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-serving-cert\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.275384 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bf8ca363-9882-4e15-bfe7-5b59686e88ea-etcd-client\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.275946 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-dir\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.276187 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-encryption-config\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.276456 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.276563 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277055 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25743472-93a4-4d4e-bf54-1d5cef4025a2-config\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277468 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277514 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277548 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277580 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277827 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.278004 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-policies\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.277127 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e041f6-1d07-4dfe-a844-ed090a258aec-config\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.278320 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b50783dc-85ca-4dd8-9ad7-08224cf4934f-serving-cert\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.282124 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6dd8d948-2f17-4ef8-8b31-572f08420bb2-serving-cert\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.283486 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-g9csq"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.284169 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.284776 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/89e148da-1707-4caa-9e28-7976df0303a5-etcd-client\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.284819 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.285940 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.287558 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.288077 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crn6w"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.289341 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q4sb7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.290643 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.291775 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.292860 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.294376 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jjm5p"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.295533 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qz8gl"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.296367 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.303923 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qz8gl"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.306043 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kbnkl"] Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.306616 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.307115 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.365628 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.365956 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1880937-0dcd-437c-8993-021954821db3-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.365982 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1880937-0dcd-437c-8993-021954821db3-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.366035 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqjdr\" (UniqueName: \"kubernetes.io/projected/f53d1c26-cef7-4963-aebc-bb7d8489f070-kube-api-access-pqjdr\") pod \"dns-operator-744455d44c-7l82d\" (UID: \"f53d1c26-cef7-4963-aebc-bb7d8489f070\") " pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.366108 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1880937-0dcd-437c-8993-021954821db3-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.366128 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f53d1c26-cef7-4963-aebc-bb7d8489f070-metrics-tls\") pod \"dns-operator-744455d44c-7l82d\" (UID: \"f53d1c26-cef7-4963-aebc-bb7d8489f070\") " pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.367623 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1880937-0dcd-437c-8993-021954821db3-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.368860 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f53d1c26-cef7-4963-aebc-bb7d8489f070-metrics-tls\") pod \"dns-operator-744455d44c-7l82d\" (UID: \"f53d1c26-cef7-4963-aebc-bb7d8489f070\") " pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.369338 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1880937-0dcd-437c-8993-021954821db3-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.385781 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.405706 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.425655 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.446180 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.465489 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.491795 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.506832 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.525559 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.546476 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.567053 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.586647 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.606131 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.626112 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.646033 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.667303 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.685896 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.705770 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.726942 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.746846 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.766101 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.786985 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.807932 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.826892 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.847509 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.867329 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.887239 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.907092 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.926470 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.947185 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.967297 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 06:13:31 crc kubenswrapper[4685]: I1204 06:13:31.986874 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.006472 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.027266 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.046672 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.065847 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.087182 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.106578 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.127023 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.146767 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.167059 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.186931 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.208072 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.226934 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.245356 4685 request.go:700] Waited for 1.004582051s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-controller-dockercfg-c2lfx&limit=500&resourceVersion=0 Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.247782 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.266661 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.287198 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.306880 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.327464 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.348502 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.367461 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.387168 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.406935 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.427276 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.453748 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.466542 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.487205 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.507880 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.527249 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.546736 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.566570 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.585669 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.606689 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.627124 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.647502 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.665972 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.697956 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.707032 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.726605 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.747661 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.767619 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.787631 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.806887 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.827189 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.866844 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.877927 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9cl2\" (UniqueName: \"kubernetes.io/projected/b50783dc-85ca-4dd8-9ad7-08224cf4934f-kube-api-access-k9cl2\") pod \"controller-manager-879f6c89f-2qkqs\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.887291 4685 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.907612 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.946929 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s96jb\" (UniqueName: \"kubernetes.io/projected/d3c9538c-1f5c-4afb-b3ef-5db510699c1a-kube-api-access-s96jb\") pod \"openshift-apiserver-operator-796bbdcf4f-4xtw5\" (UID: \"d3c9538c-1f5c-4afb-b3ef-5db510699c1a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.968061 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4tx8\" (UniqueName: \"kubernetes.io/projected/89e148da-1707-4caa-9e28-7976df0303a5-kube-api-access-d4tx8\") pod \"apiserver-7bbb656c7d-jmbn8\" (UID: \"89e148da-1707-4caa-9e28-7976df0303a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:32 crc kubenswrapper[4685]: I1204 06:13:32.985797 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqdqm\" (UniqueName: \"kubernetes.io/projected/00e041f6-1d07-4dfe-a844-ed090a258aec-kube-api-access-xqdqm\") pod \"authentication-operator-69f744f599-cplqg\" (UID: \"00e041f6-1d07-4dfe-a844-ed090a258aec\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.004131 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc2k7\" (UniqueName: \"kubernetes.io/projected/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-kube-api-access-rc2k7\") pod \"oauth-openshift-558db77b4-rfvkg\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.020511 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv8m9\" (UniqueName: \"kubernetes.io/projected/6dd8d948-2f17-4ef8-8b31-572f08420bb2-kube-api-access-qv8m9\") pod \"openshift-config-operator-7777fb866f-24w48\" (UID: \"6dd8d948-2f17-4ef8-8b31-572f08420bb2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.039820 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms5nw\" (UniqueName: \"kubernetes.io/projected/25743472-93a4-4d4e-bf54-1d5cef4025a2-kube-api-access-ms5nw\") pod \"machine-api-operator-5694c8668f-b9hw7\" (UID: \"25743472-93a4-4d4e-bf54-1d5cef4025a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.061822 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crx5m\" (UniqueName: \"kubernetes.io/projected/bf8ca363-9882-4e15-bfe7-5b59686e88ea-kube-api-access-crx5m\") pod \"apiserver-76f77b778f-fz7dx\" (UID: \"bf8ca363-9882-4e15-bfe7-5b59686e88ea\") " pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.065823 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.086173 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.106126 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.106171 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.119041 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.128184 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.136467 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.146868 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.147518 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.159721 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.166629 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.176970 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.187620 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.189286 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.203013 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.246278 4685 request.go:700] Waited for 1.879847096s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/serviceaccounts/dns-operator/token Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.265967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqjdr\" (UniqueName: \"kubernetes.io/projected/f53d1c26-cef7-4963-aebc-bb7d8489f070-kube-api-access-pqjdr\") pod \"dns-operator-744455d44c-7l82d\" (UID: \"f53d1c26-cef7-4963-aebc-bb7d8489f070\") " pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.290039 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1880937-0dcd-437c-8993-021954821db3-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-t27sr\" (UID: \"e1880937-0dcd-437c-8993-021954821db3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.291927 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19ba60b8-1577-47ae-a027-a02e32913c39-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.291958 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvwrb\" (UniqueName: \"kubernetes.io/projected/3e908422-dc37-44e1-abf1-d617bb1b7832-kube-api-access-xvwrb\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.291982 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-serving-cert\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292029 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-console-config\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292047 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-oauth-serving-cert\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292063 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-tls\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292077 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-service-ca\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292091 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-certificates\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292111 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/007053aa-6733-47ec-aeb6-7b6753d58c15-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2wd6d\" (UID: \"007053aa-6733-47ec-aeb6-7b6753d58c15\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292142 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e908422-dc37-44e1-abf1-d617bb1b7832-serving-cert\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292155 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-client\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292169 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/19ba60b8-1577-47ae-a027-a02e32913c39-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292205 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22360e53-a38b-4bc6-bb23-a993a434a880-serving-cert\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292221 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b717bcd6-4c16-491b-950c-ae2268f7c12a-config\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292236 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19ba60b8-1577-47ae-a027-a02e32913c39-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292251 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292267 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-oauth-config\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292293 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-trusted-ca-bundle\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292308 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b582cd3-5a8b-45ff-a6a7-05944059ae75-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p5tm9\" (UID: \"6b582cd3-5a8b-45ff-a6a7-05944059ae75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292325 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b717bcd6-4c16-491b-950c-ae2268f7c12a-trusted-ca\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292340 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-client-ca\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292355 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n59pf\" (UniqueName: \"kubernetes.io/projected/534c7808-899b-4952-9a29-b65dacb98ff0-kube-api-access-n59pf\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292390 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gdgw\" (UniqueName: \"kubernetes.io/projected/22360e53-a38b-4bc6-bb23-a993a434a880-kube-api-access-2gdgw\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292427 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-service-ca\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292455 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85tjl\" (UniqueName: \"kubernetes.io/projected/6b582cd3-5a8b-45ff-a6a7-05944059ae75-kube-api-access-85tjl\") pod \"cluster-samples-operator-665b6dd947-p5tm9\" (UID: \"6b582cd3-5a8b-45ff-a6a7-05944059ae75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292482 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q69mq\" (UniqueName: \"kubernetes.io/projected/19ba60b8-1577-47ae-a027-a02e32913c39-kube-api-access-q69mq\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292516 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292531 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-ca\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-config\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292564 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292632 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-bound-sa-token\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292647 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjc4d\" (UniqueName: \"kubernetes.io/projected/b717bcd6-4c16-491b-950c-ae2268f7c12a-kube-api-access-jjc4d\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292664 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np5xq\" (UniqueName: \"kubernetes.io/projected/007053aa-6733-47ec-aeb6-7b6753d58c15-kube-api-access-np5xq\") pod \"control-plane-machine-set-operator-78cbb6b69f-2wd6d\" (UID: \"007053aa-6733-47ec-aeb6-7b6753d58c15\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292679 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-config\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292695 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b717bcd6-4c16-491b-950c-ae2268f7c12a-serving-cert\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292719 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-trusted-ca\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292733 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6l48\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-kube-api-access-j6l48\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.292751 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn77j\" (UniqueName: \"kubernetes.io/projected/dbd5e94d-05ab-4ac0-9846-e700839cce2a-kube-api-access-rn77j\") pod \"downloads-7954f5f757-vkf95\" (UID: \"dbd5e94d-05ab-4ac0-9846-e700839cce2a\") " pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.293850 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:33.793837494 +0000 UTC m=+150.982068369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.326256 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.339518 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fz7dx"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.362064 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.393836 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.393999 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:33.893976642 +0000 UTC m=+151.082207427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394193 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7262de6d-5876-4f63-950c-21c7e5deaf55-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394283 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-ca\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394309 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4hbt\" (UniqueName: \"kubernetes.io/projected/2e3c86c7-f36b-430d-9950-9542224eeee9-kube-api-access-f4hbt\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394334 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/452e5252-d838-4a36-96f4-f75ca800919f-srv-cert\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394382 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-mountpoint-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394444 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55f07a2-0741-486a-8e5a-63c1b823e443-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394468 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d290da7-657f-42e8-873f-74879e993568-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394493 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjc4d\" (UniqueName: \"kubernetes.io/projected/b717bcd6-4c16-491b-950c-ae2268f7c12a-kube-api-access-jjc4d\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394517 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-socket-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394543 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np5xq\" (UniqueName: \"kubernetes.io/projected/007053aa-6733-47ec-aeb6-7b6753d58c15-kube-api-access-np5xq\") pod \"control-plane-machine-set-operator-78cbb6b69f-2wd6d\" (UID: \"007053aa-6733-47ec-aeb6-7b6753d58c15\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394566 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-config\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394586 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b717bcd6-4c16-491b-950c-ae2268f7c12a-serving-cert\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394610 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cac09002-a157-4591-94ea-6856f0888f16-auth-proxy-config\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394681 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d290da7-657f-42e8-873f-74879e993568-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394711 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6l48\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-kube-api-access-j6l48\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.394733 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/530a634c-fa68-456c-afbc-45f5a15151f1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.395072 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-ca\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396526 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-config\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396558 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-csi-data-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396608 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm82f\" (UniqueName: \"kubernetes.io/projected/9f122542-8cb9-494b-bf1b-fcae0350553a-kube-api-access-cm82f\") pod \"package-server-manager-789f6589d5-vrgc7\" (UID: \"9f122542-8cb9-494b-bf1b-fcae0350553a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396636 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-tmpfs\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396664 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2e3c86c7-f36b-430d-9950-9542224eeee9-metrics-tls\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396690 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-plugins-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396719 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvwrb\" (UniqueName: \"kubernetes.io/projected/3e908422-dc37-44e1-abf1-d617bb1b7832-kube-api-access-xvwrb\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396766 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19ba60b8-1577-47ae-a027-a02e32913c39-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-serving-cert\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396813 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtkp6\" (UniqueName: \"kubernetes.io/projected/c36c0f14-f994-4969-bb8e-f5d6e7be4bf2-kube-api-access-gtkp6\") pod \"multus-admission-controller-857f4d67dd-8ndqp\" (UID: \"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396834 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-registration-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396857 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-console-config\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396878 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-oauth-serving-cert\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396913 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-metrics-certs\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396961 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc6qv\" (UniqueName: \"kubernetes.io/projected/452e5252-d838-4a36-96f4-f75ca800919f-kube-api-access-sc6qv\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.396991 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-tls\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397009 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397028 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvbsx\" (UniqueName: \"kubernetes.io/projected/7736df92-b993-435a-ad74-4c2edd644ca3-kube-api-access-nvbsx\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397077 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/007053aa-6733-47ec-aeb6-7b6753d58c15-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2wd6d\" (UID: \"007053aa-6733-47ec-aeb6-7b6753d58c15\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397097 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e3c86c7-f36b-430d-9950-9542224eeee9-config-volume\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397115 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7b2a9d09-996b-4659-8f4f-8583055b0894-signing-key\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397136 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-client\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397183 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpklv\" (UniqueName: \"kubernetes.io/projected/7b2a9d09-996b-4659-8f4f-8583055b0894-kube-api-access-zpklv\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397201 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55269b01-b42d-4d38-8d36-2414f54adbe8-service-ca-bundle\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397223 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22360e53-a38b-4bc6-bb23-a993a434a880-serving-cert\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397239 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19ba60b8-1577-47ae-a027-a02e32913c39-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397259 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdw7h\" (UniqueName: \"kubernetes.io/projected/530a634c-fa68-456c-afbc-45f5a15151f1-kube-api-access-hdw7h\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-trusted-ca-bundle\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b582cd3-5a8b-45ff-a6a7-05944059ae75-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p5tm9\" (UID: \"6b582cd3-5a8b-45ff-a6a7-05944059ae75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397339 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cac09002-a157-4591-94ea-6856f0888f16-config\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n59pf\" (UniqueName: \"kubernetes.io/projected/534c7808-899b-4952-9a29-b65dacb98ff0-kube-api-access-n59pf\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-service-ca\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397533 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gsxv\" (UniqueName: \"kubernetes.io/projected/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-kube-api-access-5gsxv\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397553 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfq29\" (UniqueName: \"kubernetes.io/projected/a55f07a2-0741-486a-8e5a-63c1b823e443-kube-api-access-sfq29\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397620 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-apiservice-cert\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397641 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q69mq\" (UniqueName: \"kubernetes.io/projected/19ba60b8-1577-47ae-a027-a02e32913c39-kube-api-access-q69mq\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397658 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397674 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8973864-fd68-4a8b-a298-e298f515a087-cert\") pod \"ingress-canary-qz8gl\" (UID: \"e8973864-fd68-4a8b-a298-e298f515a087\") " pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397690 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e293db2-ec8b-454a-8e09-72aea57b2c97-serving-cert\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397715 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397735 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397752 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-config\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397766 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9da20a29-ca25-4f24-97ef-d081cf757bac-srv-cert\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397804 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glzpk\" (UniqueName: \"kubernetes.io/projected/cac09002-a157-4591-94ea-6856f0888f16-kube-api-access-glzpk\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397925 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc81c705-faab-44b6-8da9-96bee4bd043a-secret-volume\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397966 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmvdc\" (UniqueName: \"kubernetes.io/projected/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-kube-api-access-mmvdc\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.397985 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02ef6855-1de7-411b-9dca-e8a777d0f461-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398040 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-default-certificate\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.398055 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:33.898044937 +0000 UTC m=+151.086275712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398077 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-bound-sa-token\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398097 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-trusted-ca\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398139 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qfgd\" (UniqueName: \"kubernetes.io/projected/bfdeb965-f7e5-462b-af3e-276723bd5845-kube-api-access-2qfgd\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398159 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftvm6\" (UniqueName: \"kubernetes.io/projected/9da20a29-ca25-4f24-97ef-d081cf757bac-kube-api-access-ftvm6\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398175 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398197 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwhgf\" (UniqueName: \"kubernetes.io/projected/7262de6d-5876-4f63-950c-21c7e5deaf55-kube-api-access-kwhgf\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398214 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d290da7-657f-42e8-873f-74879e993568-config\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398231 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-trusted-ca\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398247 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/530a634c-fa68-456c-afbc-45f5a15151f1-proxy-tls\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398265 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bfdeb965-f7e5-462b-af3e-276723bd5845-certs\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398282 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhkzx\" (UniqueName: \"kubernetes.io/projected/bc81c705-faab-44b6-8da9-96bee4bd043a-kube-api-access-zhkzx\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398303 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-images\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398323 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-console-config\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398333 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn77j\" (UniqueName: \"kubernetes.io/projected/dbd5e94d-05ab-4ac0-9846-e700839cce2a-kube-api-access-rn77j\") pod \"downloads-7954f5f757-vkf95\" (UID: \"dbd5e94d-05ab-4ac0-9846-e700839cce2a\") " pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398358 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdk7g\" (UniqueName: \"kubernetes.io/projected/55269b01-b42d-4d38-8d36-2414f54adbe8-kube-api-access-qdk7g\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398375 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-metrics-tls\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398390 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7b2a9d09-996b-4659-8f4f-8583055b0894-signing-cabundle\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398423 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e293db2-ec8b-454a-8e09-72aea57b2c97-config\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398621 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19ba60b8-1577-47ae-a027-a02e32913c39-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.398990 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-trusted-ca-bundle\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.399730 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-config\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400427 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-trusted-ca\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400519 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-service-ca\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400559 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f122542-8cb9-494b-bf1b-fcae0350553a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vrgc7\" (UID: \"9f122542-8cb9-494b-bf1b-fcae0350553a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400651 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-certificates\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400706 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-proxy-tls\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400816 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bfdeb965-f7e5-462b-af3e-276723bd5845-node-bootstrap-token\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400860 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwsnw\" (UniqueName: \"kubernetes.io/projected/f2e25031-7d34-4680-8bad-592c3ecb8764-kube-api-access-rwsnw\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400888 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9jr2\" (UniqueName: \"kubernetes.io/projected/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-kube-api-access-p9jr2\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.400976 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e908422-dc37-44e1-abf1-d617bb1b7832-serving-cert\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/19ba60b8-1577-47ae-a027-a02e32913c39-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401114 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-service-ca\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401138 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-stats-auth\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401181 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cac09002-a157-4591-94ea-6856f0888f16-machine-approver-tls\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401245 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-oauth-serving-cert\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401257 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b717bcd6-4c16-491b-950c-ae2268f7c12a-config\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401322 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c36c0f14-f994-4969-bb8e-f5d6e7be4bf2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8ndqp\" (UID: \"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401371 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401418 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-oauth-config\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401523 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbsfp\" (UniqueName: \"kubernetes.io/projected/e8973864-fd68-4a8b-a298-e298f515a087-kube-api-access-pbsfp\") pod \"ingress-canary-qz8gl\" (UID: \"e8973864-fd68-4a8b-a298-e298f515a087\") " pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401544 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55f07a2-0741-486a-8e5a-63c1b823e443-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401543 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e908422-dc37-44e1-abf1-d617bb1b7832-etcd-client\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401565 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc81c705-faab-44b6-8da9-96bee4bd043a-config-volume\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401600 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef6855-1de7-411b-9dca-e8a777d0f461-config\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.401627 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b717bcd6-4c16-491b-950c-ae2268f7c12a-trusted-ca\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.403392 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-client-ca\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.410830 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gdgw\" (UniqueName: \"kubernetes.io/projected/22360e53-a38b-4bc6-bb23-a993a434a880-kube-api-access-2gdgw\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.410628 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-client-ca\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.410067 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b582cd3-5a8b-45ff-a6a7-05944059ae75-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p5tm9\" (UID: \"6b582cd3-5a8b-45ff-a6a7-05944059ae75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411132 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85tjl\" (UniqueName: \"kubernetes.io/projected/6b582cd3-5a8b-45ff-a6a7-05944059ae75-kube-api-access-85tjl\") pod \"cluster-samples-operator-665b6dd947-p5tm9\" (UID: \"6b582cd3-5a8b-45ff-a6a7-05944059ae75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411159 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrw4v\" (UniqueName: \"kubernetes.io/projected/77c35356-f854-4d2d-a4d3-44c107d6f3bd-kube-api-access-jrw4v\") pod \"migrator-59844c95c7-tgb7c\" (UID: \"77c35356-f854-4d2d-a4d3-44c107d6f3bd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411210 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-webhook-cert\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411238 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02ef6855-1de7-411b-9dca-e8a777d0f461-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411697 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-serving-cert\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411746 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22360e53-a38b-4bc6-bb23-a993a434a880-serving-cert\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411784 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/452e5252-d838-4a36-96f4-f75ca800919f-profile-collector-cert\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411804 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9da20a29-ca25-4f24-97ef-d081cf757bac-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411842 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7262de6d-5876-4f63-950c-21c7e5deaf55-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411882 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfc9z\" (UniqueName: \"kubernetes.io/projected/8e293db2-ec8b-454a-8e09-72aea57b2c97-kube-api-access-pfc9z\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.411932 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b717bcd6-4c16-491b-950c-ae2268f7c12a-serving-cert\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.413213 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/19ba60b8-1577-47ae-a027-a02e32913c39-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.413726 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e908422-dc37-44e1-abf1-d617bb1b7832-serving-cert\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.413843 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.415442 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-certificates\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.417452 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b717bcd6-4c16-491b-950c-ae2268f7c12a-trusted-ca\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.420299 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/007053aa-6733-47ec-aeb6-7b6753d58c15-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2wd6d\" (UID: \"007053aa-6733-47ec-aeb6-7b6753d58c15\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.420902 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-oauth-config\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.421943 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b717bcd6-4c16-491b-950c-ae2268f7c12a-config\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.422782 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-tls\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.427455 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.429357 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-service-ca\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.438460 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjc4d\" (UniqueName: \"kubernetes.io/projected/b717bcd6-4c16-491b-950c-ae2268f7c12a-kube-api-access-jjc4d\") pod \"console-operator-58897d9998-s6zxb\" (UID: \"b717bcd6-4c16-491b-950c-ae2268f7c12a\") " pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.448982 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6l48\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-kube-api-access-j6l48\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.463957 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np5xq\" (UniqueName: \"kubernetes.io/projected/007053aa-6733-47ec-aeb6-7b6753d58c15-kube-api-access-np5xq\") pod \"control-plane-machine-set-operator-78cbb6b69f-2wd6d\" (UID: \"007053aa-6733-47ec-aeb6-7b6753d58c15\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.482062 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvwrb\" (UniqueName: \"kubernetes.io/projected/3e908422-dc37-44e1-abf1-d617bb1b7832-kube-api-access-xvwrb\") pod \"etcd-operator-b45778765-6zcm5\" (UID: \"3e908422-dc37-44e1-abf1-d617bb1b7832\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.500943 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19ba60b8-1577-47ae-a027-a02e32913c39-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512353 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.512536 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.012516461 +0000 UTC m=+151.200747236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512615 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bfdeb965-f7e5-462b-af3e-276723bd5845-certs\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512648 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhkzx\" (UniqueName: \"kubernetes.io/projected/bc81c705-faab-44b6-8da9-96bee4bd043a-kube-api-access-zhkzx\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512674 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-images\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512715 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdk7g\" (UniqueName: \"kubernetes.io/projected/55269b01-b42d-4d38-8d36-2414f54adbe8-kube-api-access-qdk7g\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512736 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-metrics-tls\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512760 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7b2a9d09-996b-4659-8f4f-8583055b0894-signing-cabundle\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512783 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e293db2-ec8b-454a-8e09-72aea57b2c97-config\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512810 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f122542-8cb9-494b-bf1b-fcae0350553a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vrgc7\" (UID: \"9f122542-8cb9-494b-bf1b-fcae0350553a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512832 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-proxy-tls\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512853 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bfdeb965-f7e5-462b-af3e-276723bd5845-node-bootstrap-token\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512878 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwsnw\" (UniqueName: \"kubernetes.io/projected/f2e25031-7d34-4680-8bad-592c3ecb8764-kube-api-access-rwsnw\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512899 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9jr2\" (UniqueName: \"kubernetes.io/projected/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-kube-api-access-p9jr2\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512926 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-stats-auth\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512948 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cac09002-a157-4591-94ea-6856f0888f16-machine-approver-tls\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512972 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c36c0f14-f994-4969-bb8e-f5d6e7be4bf2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8ndqp\" (UID: \"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.512997 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbsfp\" (UniqueName: \"kubernetes.io/projected/e8973864-fd68-4a8b-a298-e298f515a087-kube-api-access-pbsfp\") pod \"ingress-canary-qz8gl\" (UID: \"e8973864-fd68-4a8b-a298-e298f515a087\") " pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513019 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55f07a2-0741-486a-8e5a-63c1b823e443-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513040 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc81c705-faab-44b6-8da9-96bee4bd043a-config-volume\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513064 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef6855-1de7-411b-9dca-e8a777d0f461-config\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513104 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrw4v\" (UniqueName: \"kubernetes.io/projected/77c35356-f854-4d2d-a4d3-44c107d6f3bd-kube-api-access-jrw4v\") pod \"migrator-59844c95c7-tgb7c\" (UID: \"77c35356-f854-4d2d-a4d3-44c107d6f3bd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513125 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-webhook-cert\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513146 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02ef6855-1de7-411b-9dca-e8a777d0f461-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/452e5252-d838-4a36-96f4-f75ca800919f-profile-collector-cert\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513195 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9da20a29-ca25-4f24-97ef-d081cf757bac-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513220 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7262de6d-5876-4f63-950c-21c7e5deaf55-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513243 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfc9z\" (UniqueName: \"kubernetes.io/projected/8e293db2-ec8b-454a-8e09-72aea57b2c97-kube-api-access-pfc9z\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513268 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7262de6d-5876-4f63-950c-21c7e5deaf55-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513297 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4hbt\" (UniqueName: \"kubernetes.io/projected/2e3c86c7-f36b-430d-9950-9542224eeee9-kube-api-access-f4hbt\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/452e5252-d838-4a36-96f4-f75ca800919f-srv-cert\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513341 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-mountpoint-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513363 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55f07a2-0741-486a-8e5a-63c1b823e443-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513384 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d290da7-657f-42e8-873f-74879e993568-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513424 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-socket-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513454 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cac09002-a157-4591-94ea-6856f0888f16-auth-proxy-config\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d290da7-657f-42e8-873f-74879e993568-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513500 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/530a634c-fa68-456c-afbc-45f5a15151f1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513523 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-csi-data-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513546 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm82f\" (UniqueName: \"kubernetes.io/projected/9f122542-8cb9-494b-bf1b-fcae0350553a-kube-api-access-cm82f\") pod \"package-server-manager-789f6589d5-vrgc7\" (UID: \"9f122542-8cb9-494b-bf1b-fcae0350553a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513569 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-tmpfs\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513591 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2e3c86c7-f36b-430d-9950-9542224eeee9-metrics-tls\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513614 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-plugins-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513643 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtkp6\" (UniqueName: \"kubernetes.io/projected/c36c0f14-f994-4969-bb8e-f5d6e7be4bf2-kube-api-access-gtkp6\") pod \"multus-admission-controller-857f4d67dd-8ndqp\" (UID: \"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513665 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-registration-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513691 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-metrics-certs\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513715 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc6qv\" (UniqueName: \"kubernetes.io/projected/452e5252-d838-4a36-96f4-f75ca800919f-kube-api-access-sc6qv\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513726 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e293db2-ec8b-454a-8e09-72aea57b2c97-config\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515104 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-images\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515321 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7b2a9d09-996b-4659-8f4f-8583055b0894-signing-cabundle\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.513742 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515372 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvbsx\" (UniqueName: \"kubernetes.io/projected/7736df92-b993-435a-ad74-4c2edd644ca3-kube-api-access-nvbsx\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515393 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e3c86c7-f36b-430d-9950-9542224eeee9-config-volume\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515426 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7b2a9d09-996b-4659-8f4f-8583055b0894-signing-key\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515448 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpklv\" (UniqueName: \"kubernetes.io/projected/7b2a9d09-996b-4659-8f4f-8583055b0894-kube-api-access-zpklv\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515472 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55269b01-b42d-4d38-8d36-2414f54adbe8-service-ca-bundle\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdw7h\" (UniqueName: \"kubernetes.io/projected/530a634c-fa68-456c-afbc-45f5a15151f1-kube-api-access-hdw7h\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515522 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cac09002-a157-4591-94ea-6856f0888f16-config\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515571 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gsxv\" (UniqueName: \"kubernetes.io/projected/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-kube-api-access-5gsxv\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515594 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfq29\" (UniqueName: \"kubernetes.io/projected/a55f07a2-0741-486a-8e5a-63c1b823e443-kube-api-access-sfq29\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515617 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-apiservice-cert\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515644 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515663 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8973864-fd68-4a8b-a298-e298f515a087-cert\") pod \"ingress-canary-qz8gl\" (UID: \"e8973864-fd68-4a8b-a298-e298f515a087\") " pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515684 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e293db2-ec8b-454a-8e09-72aea57b2c97-serving-cert\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515704 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515727 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9da20a29-ca25-4f24-97ef-d081cf757bac-srv-cert\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515751 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515774 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glzpk\" (UniqueName: \"kubernetes.io/projected/cac09002-a157-4591-94ea-6856f0888f16-kube-api-access-glzpk\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515794 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc81c705-faab-44b6-8da9-96bee4bd043a-secret-volume\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515816 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmvdc\" (UniqueName: \"kubernetes.io/projected/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-kube-api-access-mmvdc\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515837 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02ef6855-1de7-411b-9dca-e8a777d0f461-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515881 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-default-certificate\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515883 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515911 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-trusted-ca\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515942 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qfgd\" (UniqueName: \"kubernetes.io/projected/bfdeb965-f7e5-462b-af3e-276723bd5845-kube-api-access-2qfgd\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515963 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftvm6\" (UniqueName: \"kubernetes.io/projected/9da20a29-ca25-4f24-97ef-d081cf757bac-kube-api-access-ftvm6\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.515984 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.516007 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwhgf\" (UniqueName: \"kubernetes.io/projected/7262de6d-5876-4f63-950c-21c7e5deaf55-kube-api-access-kwhgf\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.516028 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d290da7-657f-42e8-873f-74879e993568-config\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.516068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/530a634c-fa68-456c-afbc-45f5a15151f1-proxy-tls\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.517042 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc81c705-faab-44b6-8da9-96bee4bd043a-config-volume\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.517487 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ef6855-1de7-411b-9dca-e8a777d0f461-config\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.517683 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e3c86c7-f36b-430d-9950-9542224eeee9-config-volume\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.517727 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55f07a2-0741-486a-8e5a-63c1b823e443-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.518476 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f122542-8cb9-494b-bf1b-fcae0350553a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vrgc7\" (UID: \"9f122542-8cb9-494b-bf1b-fcae0350553a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.518502 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.018487429 +0000 UTC m=+151.206718274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.519001 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bfdeb965-f7e5-462b-af3e-276723bd5845-certs\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.519930 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/530a634c-fa68-456c-afbc-45f5a15151f1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.519969 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-socket-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.520996 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7262de6d-5876-4f63-950c-21c7e5deaf55-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.521095 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7262de6d-5876-4f63-950c-21c7e5deaf55-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.521540 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-plugins-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.521581 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bfdeb965-f7e5-462b-af3e-276723bd5845-node-bootstrap-token\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.521622 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-csi-data-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.522024 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-tmpfs\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.522118 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d290da7-657f-42e8-873f-74879e993568-config\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.522237 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-registration-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.522724 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f2e25031-7d34-4680-8bad-592c3ecb8764-mountpoint-dir\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.523251 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-trusted-ca\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.524107 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-metrics-tls\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.525922 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d290da7-657f-42e8-873f-74879e993568-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.528490 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.530209 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55269b01-b42d-4d38-8d36-2414f54adbe8-service-ca-bundle\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.530350 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c36c0f14-f994-4969-bb8e-f5d6e7be4bf2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8ndqp\" (UID: \"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.530593 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-stats-auth\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.531102 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/530a634c-fa68-456c-afbc-45f5a15151f1-proxy-tls\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.531349 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55f07a2-0741-486a-8e5a-63c1b823e443-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.531583 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9da20a29-ca25-4f24-97ef-d081cf757bac-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.532016 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2e3c86c7-f36b-430d-9950-9542224eeee9-metrics-tls\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.532242 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/452e5252-d838-4a36-96f4-f75ca800919f-profile-collector-cert\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.532466 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8973864-fd68-4a8b-a298-e298f515a087-cert\") pod \"ingress-canary-qz8gl\" (UID: \"e8973864-fd68-4a8b-a298-e298f515a087\") " pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.532798 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7b2a9d09-996b-4659-8f4f-8583055b0894-signing-key\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.533725 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-apiservice-cert\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.534572 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-default-certificate\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.534654 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/452e5252-d838-4a36-96f4-f75ca800919f-srv-cert\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.534915 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.535080 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-webhook-cert\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.535140 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02ef6855-1de7-411b-9dca-e8a777d0f461-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.535557 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e293db2-ec8b-454a-8e09-72aea57b2c97-serving-cert\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.536204 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55269b01-b42d-4d38-8d36-2414f54adbe8-metrics-certs\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.536229 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9da20a29-ca25-4f24-97ef-d081cf757bac-srv-cert\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.536657 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc81c705-faab-44b6-8da9-96bee4bd043a-secret-volume\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.536981 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-proxy-tls\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.537274 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cac09002-a157-4591-94ea-6856f0888f16-config\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.537353 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cac09002-a157-4591-94ea-6856f0888f16-auth-proxy-config\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.537503 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cac09002-a157-4591-94ea-6856f0888f16-machine-approver-tls\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.544420 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rfvkg"] Dec 04 06:13:33 crc kubenswrapper[4685]: W1204 06:13:33.550993 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ae778b1_6a99_4fec_8a14_d6d2322f18e7.slice/crio-1dba8640be43b88ef2f5d02fda9d67eec072e49a48ca4b1eedc54918b95cf09c WatchSource:0}: Error finding container 1dba8640be43b88ef2f5d02fda9d67eec072e49a48ca4b1eedc54918b95cf09c: Status 404 returned error can't find the container with id 1dba8640be43b88ef2f5d02fda9d67eec072e49a48ca4b1eedc54918b95cf09c Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.551794 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-bound-sa-token\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.562624 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q69mq\" (UniqueName: \"kubernetes.io/projected/19ba60b8-1577-47ae-a027-a02e32913c39-kube-api-access-q69mq\") pod \"cluster-image-registry-operator-dc59b4c8b-ls5hs\" (UID: \"19ba60b8-1577-47ae-a027-a02e32913c39\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.565557 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7l82d"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.575034 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qkqs"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.581803 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn77j\" (UniqueName: \"kubernetes.io/projected/dbd5e94d-05ab-4ac0-9846-e700839cce2a-kube-api-access-rn77j\") pod \"downloads-7954f5f757-vkf95\" (UID: \"dbd5e94d-05ab-4ac0-9846-e700839cce2a\") " pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.586629 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.588192 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8"] Dec 04 06:13:33 crc kubenswrapper[4685]: W1204 06:13:33.599494 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb50783dc_85ca_4dd8_9ad7_08224cf4934f.slice/crio-e2d926136d27817814b36d6ec5f8a3b6dc40647513ebcaa248553a9245a2c127 WatchSource:0}: Error finding container e2d926136d27817814b36d6ec5f8a3b6dc40647513ebcaa248553a9245a2c127: Status 404 returned error can't find the container with id e2d926136d27817814b36d6ec5f8a3b6dc40647513ebcaa248553a9245a2c127 Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.609578 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n59pf\" (UniqueName: \"kubernetes.io/projected/534c7808-899b-4952-9a29-b65dacb98ff0-kube-api-access-n59pf\") pod \"console-f9d7485db-s6kbn\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.614767 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.616724 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.616852 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.116828918 +0000 UTC m=+151.305059693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.617101 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.617545 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr"] Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.618800 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.118790042 +0000 UTC m=+151.307020817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.621815 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gdgw\" (UniqueName: \"kubernetes.io/projected/22360e53-a38b-4bc6-bb23-a993a434a880-kube-api-access-2gdgw\") pod \"route-controller-manager-6576b87f9c-b5sjh\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: W1204 06:13:33.626687 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1880937_0dcd_437c_8993_021954821db3.slice/crio-ee3eb546dc0fb1ceeca54a0c0662dc71c70920b384baea0ae1eb90716cc5347d WatchSource:0}: Error finding container ee3eb546dc0fb1ceeca54a0c0662dc71c70920b384baea0ae1eb90716cc5347d: Status 404 returned error can't find the container with id ee3eb546dc0fb1ceeca54a0c0662dc71c70920b384baea0ae1eb90716cc5347d Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.633755 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.644029 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85tjl\" (UniqueName: \"kubernetes.io/projected/6b582cd3-5a8b-45ff-a6a7-05944059ae75-kube-api-access-85tjl\") pod \"cluster-samples-operator-665b6dd947-p5tm9\" (UID: \"6b582cd3-5a8b-45ff-a6a7-05944059ae75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.644102 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b9hw7"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.653911 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.676746 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.682781 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdk7g\" (UniqueName: \"kubernetes.io/projected/55269b01-b42d-4d38-8d36-2414f54adbe8-kube-api-access-qdk7g\") pod \"router-default-5444994796-qxx2n\" (UID: \"55269b01-b42d-4d38-8d36-2414f54adbe8\") " pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.691483 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cplqg"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.701584 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhkzx\" (UniqueName: \"kubernetes.io/projected/bc81c705-faab-44b6-8da9-96bee4bd043a-kube-api-access-zhkzx\") pod \"collect-profiles-29413800-slrlm\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.711125 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-24w48"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.715996 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.719297 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.719793 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.219778209 +0000 UTC m=+151.408008984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.728554 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbsfp\" (UniqueName: \"kubernetes.io/projected/e8973864-fd68-4a8b-a298-e298f515a087-kube-api-access-pbsfp\") pod \"ingress-canary-qz8gl\" (UID: \"e8973864-fd68-4a8b-a298-e298f515a087\") " pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.732263 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.746961 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvbsx\" (UniqueName: \"kubernetes.io/projected/7736df92-b993-435a-ad74-4c2edd644ca3-kube-api-access-nvbsx\") pod \"marketplace-operator-79b997595-crn6w\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.762143 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwsnw\" (UniqueName: \"kubernetes.io/projected/f2e25031-7d34-4680-8bad-592c3ecb8764-kube-api-access-rwsnw\") pod \"csi-hostpathplugin-jjm5p\" (UID: \"f2e25031-7d34-4680-8bad-592c3ecb8764\") " pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.788423 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrw4v\" (UniqueName: \"kubernetes.io/projected/77c35356-f854-4d2d-a4d3-44c107d6f3bd-kube-api-access-jrw4v\") pod \"migrator-59844c95c7-tgb7c\" (UID: \"77c35356-f854-4d2d-a4d3-44c107d6f3bd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.806682 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9jr2\" (UniqueName: \"kubernetes.io/projected/63fa9dc9-ddb8-4731-ad67-cec8fb8a7439-kube-api-access-p9jr2\") pod \"packageserver-d55dfcdfc-lkbzh\" (UID: \"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.813850 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.820442 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.820721 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6zcm5"] Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.820828 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.320813457 +0000 UTC m=+151.509044232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.821634 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glzpk\" (UniqueName: \"kubernetes.io/projected/cac09002-a157-4591-94ea-6856f0888f16-kube-api-access-glzpk\") pod \"machine-approver-56656f9798-hwvbm\" (UID: \"cac09002-a157-4591-94ea-6856f0888f16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.824871 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.843894 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" event={"ID":"f53d1c26-cef7-4963-aebc-bb7d8489f070","Type":"ContainerStarted","Data":"47bea560709aeecf0d335e86c1213f9262c5f138bd2b5f454d225b4bbd8bb7eb"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.843954 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gsxv\" (UniqueName: \"kubernetes.io/projected/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-kube-api-access-5gsxv\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.845215 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" event={"ID":"3ae778b1-6a99-4fec-8a14-d6d2322f18e7","Type":"ContainerStarted","Data":"1dba8640be43b88ef2f5d02fda9d67eec072e49a48ca4b1eedc54918b95cf09c"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.845654 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.847217 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" event={"ID":"b50783dc-85ca-4dd8-9ad7-08224cf4934f","Type":"ContainerStarted","Data":"e2d926136d27817814b36d6ec5f8a3b6dc40647513ebcaa248553a9245a2c127"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.848029 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" event={"ID":"e1880937-0dcd-437c-8993-021954821db3","Type":"ContainerStarted","Data":"ee3eb546dc0fb1ceeca54a0c0662dc71c70920b384baea0ae1eb90716cc5347d"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.855121 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" event={"ID":"bf8ca363-9882-4e15-bfe7-5b59686e88ea","Type":"ContainerStarted","Data":"a534c8ca8fe5285b66f03f5bc82266904a2541873a3e7fe2dccdf1888e0dbd96"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.856156 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" event={"ID":"25743472-93a4-4d4e-bf54-1d5cef4025a2","Type":"ContainerStarted","Data":"afe3c25bc229ed9ed6b0ea0f446ba5e03b46606bd6f0ff7e4e73dd2222ff3660"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.857248 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" event={"ID":"89e148da-1707-4caa-9e28-7976df0303a5","Type":"ContainerStarted","Data":"1e09900065e524d6d8d50e363f6a3fe63b26d9ea034c6551c3527b56936c35cf"} Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.859505 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.860059 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qfgd\" (UniqueName: \"kubernetes.io/projected/bfdeb965-f7e5-462b-af3e-276723bd5845-kube-api-access-2qfgd\") pod \"machine-config-server-kbnkl\" (UID: \"bfdeb965-f7e5-462b-af3e-276723bd5845\") " pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.867532 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.877573 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.884466 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh"] Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.886153 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.886724 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftvm6\" (UniqueName: \"kubernetes.io/projected/9da20a29-ca25-4f24-97ef-d081cf757bac-kube-api-access-ftvm6\") pod \"olm-operator-6b444d44fb-gt6kp\" (UID: \"9da20a29-ca25-4f24-97ef-d081cf757bac\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.901833 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmvdc\" (UniqueName: \"kubernetes.io/projected/3334bba6-dbe5-4be6-909e-95cd1b26d6ce-kube-api-access-mmvdc\") pod \"machine-config-operator-74547568cd-pwfqr\" (UID: \"3334bba6-dbe5-4be6-909e-95cd1b26d6ce\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.911108 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.918251 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qz8gl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.920016 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02ef6855-1de7-411b-9dca-e8a777d0f461-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cvsf2\" (UID: \"02ef6855-1de7-411b-9dca-e8a777d0f461\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.921429 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.921609 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.421589837 +0000 UTC m=+151.609820622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.921945 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:33 crc kubenswrapper[4685]: E1204 06:13:33.922240 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.422233468 +0000 UTC m=+151.610464243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.924384 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kbnkl" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.971429 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d290da7-657f-42e8-873f-74879e993568-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pwwjm\" (UID: \"5d290da7-657f-42e8-873f-74879e993568\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:33 crc kubenswrapper[4685]: I1204 06:13:33.987808 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm82f\" (UniqueName: \"kubernetes.io/projected/9f122542-8cb9-494b-bf1b-fcae0350553a-kube-api-access-cm82f\") pod \"package-server-manager-789f6589d5-vrgc7\" (UID: \"9f122542-8cb9-494b-bf1b-fcae0350553a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.000767 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwhgf\" (UniqueName: \"kubernetes.io/projected/7262de6d-5876-4f63-950c-21c7e5deaf55-kube-api-access-kwhgf\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbt4q\" (UID: \"7262de6d-5876-4f63-950c-21c7e5deaf55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.016864 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.020313 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtkp6\" (UniqueName: \"kubernetes.io/projected/c36c0f14-f994-4969-bb8e-f5d6e7be4bf2-kube-api-access-gtkp6\") pod \"multus-admission-controller-857f4d67dd-8ndqp\" (UID: \"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.023657 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.023822 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.523802034 +0000 UTC m=+151.712032809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.024075 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.024374 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.524366623 +0000 UTC m=+151.712597398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.040712 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpklv\" (UniqueName: \"kubernetes.io/projected/7b2a9d09-996b-4659-8f4f-8583055b0894-kube-api-access-zpklv\") pod \"service-ca-9c57cc56f-q4sb7\" (UID: \"7b2a9d09-996b-4659-8f4f-8583055b0894\") " pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.045783 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.051301 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.057042 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.061292 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0c5a2fb-94ab-4f05-acba-73a19cb2e798-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6j5ws\" (UID: \"d0c5a2fb-94ab-4f05-acba-73a19cb2e798\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.065112 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.085013 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.101317 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdw7h\" (UniqueName: \"kubernetes.io/projected/530a634c-fa68-456c-afbc-45f5a15151f1-kube-api-access-hdw7h\") pod \"machine-config-controller-84d6567774-pcxg4\" (UID: \"530a634c-fa68-456c-afbc-45f5a15151f1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.104695 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.112778 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.120938 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc6qv\" (UniqueName: \"kubernetes.io/projected/452e5252-d838-4a36-96f4-f75ca800919f-kube-api-access-sc6qv\") pod \"catalog-operator-68c6474976-px7gh\" (UID: \"452e5252-d838-4a36-96f4-f75ca800919f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.121186 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.125681 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.125807 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.625776963 +0000 UTC m=+151.814007758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.126058 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.126497 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.626482976 +0000 UTC m=+151.814713801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.129996 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.136539 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.141495 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4hbt\" (UniqueName: \"kubernetes.io/projected/2e3c86c7-f36b-430d-9950-9542224eeee9-kube-api-access-f4hbt\") pod \"dns-default-g9csq\" (UID: \"2e3c86c7-f36b-430d-9950-9542224eeee9\") " pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.175003 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.227444 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.227772 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.727736632 +0000 UTC m=+151.915967457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.227938 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.228647 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.728625341 +0000 UTC m=+151.916856166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.291952 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfc9z\" (UniqueName: \"kubernetes.io/projected/8e293db2-ec8b-454a-8e09-72aea57b2c97-kube-api-access-pfc9z\") pod \"service-ca-operator-777779d784-tgrh2\" (UID: \"8e293db2-ec8b-454a-8e09-72aea57b2c97\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.296148 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfq29\" (UniqueName: \"kubernetes.io/projected/a55f07a2-0741-486a-8e5a-63c1b823e443-kube-api-access-sfq29\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvcw7\" (UID: \"a55f07a2-0741-486a-8e5a-63c1b823e443\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:34 crc kubenswrapper[4685]: W1204 06:13:34.326009 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3c9538c_1f5c_4afb_b3ef_5db510699c1a.slice/crio-599b32464b977732c1302adfc0146ae06a8c11304e0723881e3094528da5248b WatchSource:0}: Error finding container 599b32464b977732c1302adfc0146ae06a8c11304e0723881e3094528da5248b: Status 404 returned error can't find the container with id 599b32464b977732c1302adfc0146ae06a8c11304e0723881e3094528da5248b Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.328832 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.329027 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.828998748 +0000 UTC m=+152.017229533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.329102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.329500 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.829482844 +0000 UTC m=+152.017713639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.337546 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.392492 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.400327 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.429990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.430599 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:34.930565544 +0000 UTC m=+152.118796339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.452974 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.531061 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.531846 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.03183146 +0000 UTC m=+152.220062235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.632302 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.632495 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.132467735 +0000 UTC m=+152.320698510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.632913 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.633215 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.133206388 +0000 UTC m=+152.321437153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.640689 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s6zxb"] Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.668169 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q"] Dec 04 06:13:34 crc kubenswrapper[4685]: W1204 06:13:34.708802 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb717bcd6_4c16_491b_950c_ae2268f7c12a.slice/crio-ffcbc25a0fdb45addd4545233cc5d1f28a67d135d2bc18e90379770f3f0a6bae WatchSource:0}: Error finding container ffcbc25a0fdb45addd4545233cc5d1f28a67d135d2bc18e90379770f3f0a6bae: Status 404 returned error can't find the container with id ffcbc25a0fdb45addd4545233cc5d1f28a67d135d2bc18e90379770f3f0a6bae Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.737861 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.742790 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.242764549 +0000 UTC m=+152.430995324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.843698 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:34 crc kubenswrapper[4685]: E1204 06:13:34.844134 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.344117408 +0000 UTC m=+152.532348183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:34 crc kubenswrapper[4685]: I1204 06:13:34.868876 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" event={"ID":"7262de6d-5876-4f63-950c-21c7e5deaf55","Type":"ContainerStarted","Data":"c962bebf9e3574376f1ebdf2fbf325a7af2f577b32f92a35ff73a7d5daaec420"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.881169 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" event={"ID":"e1880937-0dcd-437c-8993-021954821db3","Type":"ContainerStarted","Data":"f19bb09c750b9326e642951cd7bf2aec41dcf29c8663db664743e9a3fddef62b"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.882206 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" event={"ID":"3e908422-dc37-44e1-abf1-d617bb1b7832","Type":"ContainerStarted","Data":"be556c5bcaff69b049cd0dae3d5423f8409c4f0832431efd882c3f2a3c5f8ebf"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.883301 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" event={"ID":"b717bcd6-4c16-491b-950c-ae2268f7c12a","Type":"ContainerStarted","Data":"ffcbc25a0fdb45addd4545233cc5d1f28a67d135d2bc18e90379770f3f0a6bae"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.884877 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" event={"ID":"3ae778b1-6a99-4fec-8a14-d6d2322f18e7","Type":"ContainerStarted","Data":"ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.885723 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.887112 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" event={"ID":"b50783dc-85ca-4dd8-9ad7-08224cf4934f","Type":"ContainerStarted","Data":"785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.887631 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.888315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" event={"ID":"cac09002-a157-4591-94ea-6856f0888f16","Type":"ContainerStarted","Data":"cd228a64dcf47746131dc318bb2fb9676cef28021ba1ec5e3aa372525e772e16"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.891311 4685 generic.go:334] "Generic (PLEG): container finished" podID="bf8ca363-9882-4e15-bfe7-5b59686e88ea" containerID="39761621590d092fc9959b00bbde6837bc6e54e6b0e76308ad84b1e0ff0ff331" exitCode=0 Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:34.891351 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" event={"ID":"bf8ca363-9882-4e15-bfe7-5b59686e88ea","Type":"ContainerDied","Data":"39761621590d092fc9959b00bbde6837bc6e54e6b0e76308ad84b1e0ff0ff331"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.015683 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.016081 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.516063566 +0000 UTC m=+152.704294341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.034367 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" event={"ID":"22360e53-a38b-4bc6-bb23-a993a434a880","Type":"ContainerStarted","Data":"e5fe358dde70a71aa1cf753d94fd2d232c0f044f29eedf37fa6060dae1a7928d"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.037108 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" event={"ID":"f53d1c26-cef7-4963-aebc-bb7d8489f070","Type":"ContainerStarted","Data":"3e31194ef99ac00ea92bd7c997e45e618f4a46a880de5a7779c18d38b85b8ca7"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.039395 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" event={"ID":"00e041f6-1d07-4dfe-a844-ed090a258aec","Type":"ContainerStarted","Data":"6194cf7bbe8ad595ff0d9b3e49a621ebdb0b5c0b4a730381b4238937d01726d6"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.053289 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" event={"ID":"d3c9538c-1f5c-4afb-b3ef-5db510699c1a","Type":"ContainerStarted","Data":"599b32464b977732c1302adfc0146ae06a8c11304e0723881e3094528da5248b"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.065273 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" event={"ID":"6dd8d948-2f17-4ef8-8b31-572f08420bb2","Type":"ContainerStarted","Data":"3b26bbf1b06d11976b5ac3143e27402bb2119e9f83e3a7dba7673cf3f5f97975"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.065362 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.068368 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" event={"ID":"19ba60b8-1577-47ae-a027-a02e32913c39","Type":"ContainerStarted","Data":"a79c5fa5b40669a3c2d5f7de8e393836bff9db50d9693a24323decd658cc03fb"} Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.111343 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d"] Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.117164 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.118733 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.618719798 +0000 UTC m=+152.806950573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.199440 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.223809 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.223885 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.723869403 +0000 UTC m=+152.912100178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.224326 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.224690 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.724682489 +0000 UTC m=+152.912913264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.328772 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.329519 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.829480542 +0000 UTC m=+153.017711317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.430855 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.431257 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:35.931244224 +0000 UTC m=+153.119474999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: W1204 06:13:35.498018 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod007053aa_6733_47ec_aeb6_7b6753d58c15.slice/crio-64f3190916849c5e676a94e6a72cfc1f97a6e1283479d98c410e97bc8f965587 WatchSource:0}: Error finding container 64f3190916849c5e676a94e6a72cfc1f97a6e1283479d98c410e97bc8f965587: Status 404 returned error can't find the container with id 64f3190916849c5e676a94e6a72cfc1f97a6e1283479d98c410e97bc8f965587 Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.534516 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.534800 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.034783525 +0000 UTC m=+153.223014290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.632846 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.643259 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.643659 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.143643704 +0000 UTC m=+153.331874479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.703736 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9"] Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.727379 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-g9csq"] Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.743823 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.744270 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.244251738 +0000 UTC m=+153.432482513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.849954 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.850377 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.350363534 +0000 UTC m=+153.538594309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:35 crc kubenswrapper[4685]: I1204 06:13:35.954825 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:35 crc kubenswrapper[4685]: E1204 06:13:35.955240 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.455223938 +0000 UTC m=+153.643454713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.003954 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" podStartSLOduration=123.003938183 podStartE2EDuration="2m3.003938183s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.002706062 +0000 UTC m=+153.190936837" watchObservedRunningTime="2025-12-04 06:13:36.003938183 +0000 UTC m=+153.192168958" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.057908 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.058330 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.558315745 +0000 UTC m=+153.746546530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.092057 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" podStartSLOduration=123.092043132 podStartE2EDuration="2m3.092043132s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.089982224 +0000 UTC m=+153.278212999" watchObservedRunningTime="2025-12-04 06:13:36.092043132 +0000 UTC m=+153.280273907" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.095708 4685 generic.go:334] "Generic (PLEG): container finished" podID="6dd8d948-2f17-4ef8-8b31-572f08420bb2" containerID="8077125b901ea78967f18e329bdd1a162fcf5898acf6168e3d14329f7540f261" exitCode=0 Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.095775 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" event={"ID":"6dd8d948-2f17-4ef8-8b31-572f08420bb2","Type":"ContainerDied","Data":"8077125b901ea78967f18e329bdd1a162fcf5898acf6168e3d14329f7540f261"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.109144 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" event={"ID":"25743472-93a4-4d4e-bf54-1d5cef4025a2","Type":"ContainerStarted","Data":"5c9c6b38ed53a9ad5f1c06d6276e885aab07ac98e55a5faf168b79e5d8343d86"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.112905 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kbnkl" event={"ID":"bfdeb965-f7e5-462b-af3e-276723bd5845","Type":"ContainerStarted","Data":"d86a311e561a8eab025c837d7a62b99d593a9738206c20ed9bf1cdb00c93dfd6"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.131567 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.131835 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.143784 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s6kbn"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.152694 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crn6w"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.157357 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.158633 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.158837 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.158886 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.158905 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.158928 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.160502 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" event={"ID":"b717bcd6-4c16-491b-950c-ae2268f7c12a","Type":"ContainerStarted","Data":"2ec92f350334af77a8eaafe8422988c3fca092996c4e80b8afb2afa6ba07a6b0"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.161524 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.167436 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.66739992 +0000 UTC m=+153.855630695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.169095 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-qxx2n" event={"ID":"55269b01-b42d-4d38-8d36-2414f54adbe8","Type":"ContainerStarted","Data":"a443176a8f5a3cd9393a1e968b642368167c3d7457dd1465243ebbb13b91201e"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.173133 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.185898 4685 patch_prober.go:28] interesting pod/console-operator-58897d9998-s6zxb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.185964 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" podUID="b717bcd6-4c16-491b-950c-ae2268f7c12a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.187312 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.187379 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.189513 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.189833 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" podStartSLOduration=123.189812302 podStartE2EDuration="2m3.189812302s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.186466422 +0000 UTC m=+153.374697217" watchObservedRunningTime="2025-12-04 06:13:36.189812302 +0000 UTC m=+153.378043107" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.227584 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" event={"ID":"00e041f6-1d07-4dfe-a844-ed090a258aec","Type":"ContainerStarted","Data":"583984710c9649cee1ca36fe525b492e41860aa4f2b125996ae339e93ee8be63"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.259790 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" event={"ID":"d3c9538c-1f5c-4afb-b3ef-5db510699c1a","Type":"ContainerStarted","Data":"90a7c425f1456d8db4cc9391d3b6bf0baf4ace6bc85974de920206c99aba1ca8"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.268041 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.268389 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.768374556 +0000 UTC m=+153.956605331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.270740 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-g9csq" event={"ID":"2e3c86c7-f36b-430d-9950-9542224eeee9","Type":"ContainerStarted","Data":"2683d8264f9516f55a32168b15e37a1b475bc8189ae4d6d4ebeddbb22fdc0fca"} Dec 04 06:13:36 crc kubenswrapper[4685]: W1204 06:13:36.318093 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod534c7808_899b_4952_9a29_b65dacb98ff0.slice/crio-dbaf752c85d44102e4754c1948f6d24947e2e8c533c2373f40d0148de00348a2 WatchSource:0}: Error finding container dbaf752c85d44102e4754c1948f6d24947e2e8c533c2373f40d0148de00348a2: Status 404 returned error can't find the container with id dbaf752c85d44102e4754c1948f6d24947e2e8c533c2373f40d0148de00348a2 Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.331827 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" event={"ID":"22360e53-a38b-4bc6-bb23-a993a434a880","Type":"ContainerStarted","Data":"838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.333296 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.355368 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-cplqg" podStartSLOduration=123.355352148 podStartE2EDuration="2m3.355352148s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.323978919 +0000 UTC m=+153.512209694" watchObservedRunningTime="2025-12-04 06:13:36.355352148 +0000 UTC m=+153.543582923" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.370996 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.372710 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:36.872688983 +0000 UTC m=+154.060919768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.383577 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" podStartSLOduration=122.383562123 podStartE2EDuration="2m2.383562123s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.381507985 +0000 UTC m=+153.569738760" watchObservedRunningTime="2025-12-04 06:13:36.383562123 +0000 UTC m=+153.571792898" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.383851 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4xtw5" podStartSLOduration=123.383846533 podStartE2EDuration="2m3.383846533s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.355527954 +0000 UTC m=+153.543758729" watchObservedRunningTime="2025-12-04 06:13:36.383846533 +0000 UTC m=+153.572077308" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.399225 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" event={"ID":"007053aa-6733-47ec-aeb6-7b6753d58c15","Type":"ContainerStarted","Data":"64f3190916849c5e676a94e6a72cfc1f97a6e1283479d98c410e97bc8f965587"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.413931 4685 generic.go:334] "Generic (PLEG): container finished" podID="89e148da-1707-4caa-9e28-7976df0303a5" containerID="31442eaa3978ee036f7e9322961d0704dbf72d9e8ca24a438a0321ae48eab50d" exitCode=0 Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.414034 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" event={"ID":"89e148da-1707-4caa-9e28-7976df0303a5","Type":"ContainerDied","Data":"31442eaa3978ee036f7e9322961d0704dbf72d9e8ca24a438a0321ae48eab50d"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.448760 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" event={"ID":"19ba60b8-1577-47ae-a027-a02e32913c39","Type":"ContainerStarted","Data":"d77f2d5312154362b545f7afaaaff947ceba0843cba92425ac9d67c6d60f3c03"} Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.452544 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.473234 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.491024 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.492043 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.493192 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-t27sr" podStartSLOduration=122.493172535 podStartE2EDuration="2m2.493172535s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.492528844 +0000 UTC m=+153.680759619" watchObservedRunningTime="2025-12-04 06:13:36.493172535 +0000 UTC m=+153.681403310" Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.500981 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.000954873 +0000 UTC m=+154.189185768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.537192 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q4sb7"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.538998 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ls5hs" podStartSLOduration=123.538965853 podStartE2EDuration="2m3.538965853s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:36.520402418 +0000 UTC m=+153.708633193" watchObservedRunningTime="2025-12-04 06:13:36.538965853 +0000 UTC m=+153.727196628" Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.593999 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.594437 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.09439494 +0000 UTC m=+154.282625715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.606682 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.633222 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.648328 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jjm5p"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.680715 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.696320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.696691 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.196673069 +0000 UTC m=+154.384903844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.711127 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.713743 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.792042 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.801755 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.802023 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.30200795 +0000 UTC m=+154.490238725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.807664 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.809807 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vkf95"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.831380 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7"] Dec 04 06:13:36 crc kubenswrapper[4685]: I1204 06:13:36.903522 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:36 crc kubenswrapper[4685]: E1204 06:13:36.903880 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.403855975 +0000 UTC m=+154.592086740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.007701 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.008169 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.508149321 +0000 UTC m=+154.696380096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.008677 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.009192 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.509183286 +0000 UTC m=+154.697414051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.034497 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4"] Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.036287 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8ndqp"] Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.110131 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2"] Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.110401 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c"] Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.117093 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.123510 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.623486493 +0000 UTC m=+154.811717268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.191469 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.191503 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7"] Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.191524 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qz8gl"] Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.221612 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.221925 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.721912445 +0000 UTC m=+154.910143220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.324748 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.325598 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.825582461 +0000 UTC m=+155.013813236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.426949 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.427749 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:37.927735546 +0000 UTC m=+155.115966321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: W1204 06:13:37.479546 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-735a58d52edf84d0ed6635f8e8bac74d94e9827488cf7cfa5210d5a92d35607d WatchSource:0}: Error finding container 735a58d52edf84d0ed6635f8e8bac74d94e9827488cf7cfa5210d5a92d35607d: Status 404 returned error can't find the container with id 735a58d52edf84d0ed6635f8e8bac74d94e9827488cf7cfa5210d5a92d35607d Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.531221 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.531599 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.031580147 +0000 UTC m=+155.219810922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.538957 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" event={"ID":"cac09002-a157-4591-94ea-6856f0888f16","Type":"ContainerStarted","Data":"00458978c10d086ce255bffc71a62345eb6d234e3b2c821d3a1df7955cb3477c"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.539000 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" event={"ID":"cac09002-a157-4591-94ea-6856f0888f16","Type":"ContainerStarted","Data":"b4ab29b47d83b906d6512f04680d2aba3bb6183fe854a0b0ca4537f5f9a2a525"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.541817 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" event={"ID":"bf8ca363-9882-4e15-bfe7-5b59686e88ea","Type":"ContainerStarted","Data":"f0c89ca2d5d5e379e5c91abd26f4c15ad247d61c4ffe63e218262220b5b7b6b9"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.541839 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" event={"ID":"bf8ca363-9882-4e15-bfe7-5b59686e88ea","Type":"ContainerStarted","Data":"73a46e8d925a09aee68bcd82a808afdbaa5fb880a2e2ec9a42b8237fc689924c"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.563328 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwvbm" podStartSLOduration=124.563307899 podStartE2EDuration="2m4.563307899s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:37.561039834 +0000 UTC m=+154.749270609" watchObservedRunningTime="2025-12-04 06:13:37.563307899 +0000 UTC m=+154.751538674" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.575535 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" event={"ID":"7262de6d-5876-4f63-950c-21c7e5deaf55","Type":"ContainerStarted","Data":"ec9617e772b7a78459a84e72adcf7c0cf8929eafd1edc0b113e69a0aa5d82c14"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.586619 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" event={"ID":"6b582cd3-5a8b-45ff-a6a7-05944059ae75","Type":"ContainerStarted","Data":"a0e98bb2acd2fbb017312b413c37ef2eca38934bab3a9c604a11125c505b89b4"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.586669 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" event={"ID":"6b582cd3-5a8b-45ff-a6a7-05944059ae75","Type":"ContainerStarted","Data":"8d96140d267aa76d4877c30746ade82559a5266301cc6474c660a64ece8963d2"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.592046 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"eeac8e0f956c729c3ad23d460ed962a93a019c698672e42dbba0e970803c8030"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.606896 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qz8gl" event={"ID":"e8973864-fd68-4a8b-a298-e298f515a087","Type":"ContainerStarted","Data":"16323ed0aa7c2b40fd31e6b001e0fe83bb81ad11b2b9a5f41a1ba80341efe8e5"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.620068 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" event={"ID":"02ef6855-1de7-411b-9dca-e8a777d0f461","Type":"ContainerStarted","Data":"be50cf4d73c0d85614cae30b88ba06829f13fb83ab8695960917611e3bab5164"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.624541 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" podStartSLOduration=124.624495306 podStartE2EDuration="2m4.624495306s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:37.620299178 +0000 UTC m=+154.808529963" watchObservedRunningTime="2025-12-04 06:13:37.624495306 +0000 UTC m=+154.812726081" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.625160 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" event={"ID":"8e293db2-ec8b-454a-8e09-72aea57b2c97","Type":"ContainerStarted","Data":"d3459b8051aa2225adab834730f002e605bb95122961a55d9c17fdda82d02e8c"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.634007 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.640427 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.140378492 +0000 UTC m=+155.328609267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.648642 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kbnkl" event={"ID":"bfdeb965-f7e5-462b-af3e-276723bd5845","Type":"ContainerStarted","Data":"dbe8dc64d37de3bb467e33df58eaf37cbc02e56b1727186672c37c95fb0ed1ee"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.649955 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbt4q" podStartSLOduration=124.649930949 podStartE2EDuration="2m4.649930949s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:37.648869034 +0000 UTC m=+154.837099809" watchObservedRunningTime="2025-12-04 06:13:37.649930949 +0000 UTC m=+154.838161724" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.702234 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kbnkl" podStartSLOduration=6.702214391 podStartE2EDuration="6.702214391s" podCreationTimestamp="2025-12-04 06:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:37.701813578 +0000 UTC m=+154.890044353" watchObservedRunningTime="2025-12-04 06:13:37.702214391 +0000 UTC m=+154.890445166" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.738509 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.739761 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.239742455 +0000 UTC m=+155.427973230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.765703 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" event={"ID":"530a634c-fa68-456c-afbc-45f5a15151f1","Type":"ContainerStarted","Data":"8cd50c7a362df5a01f04e414926a3609440a776bc45b0bbdc831fc11506d9b87"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.788184 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6kbn" event={"ID":"534c7808-899b-4952-9a29-b65dacb98ff0","Type":"ContainerStarted","Data":"f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.788481 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6kbn" event={"ID":"534c7808-899b-4952-9a29-b65dacb98ff0","Type":"ContainerStarted","Data":"dbaf752c85d44102e4754c1948f6d24947e2e8c533c2373f40d0148de00348a2"} Dec 04 06:13:37 crc kubenswrapper[4685]: W1204 06:13:37.822118 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-038f171d600d493add18d92fee2bc0ed4fa96c59c9dcd002a109d1aa872acaf8 WatchSource:0}: Error finding container 038f171d600d493add18d92fee2bc0ed4fa96c59c9dcd002a109d1aa872acaf8: Status 404 returned error can't find the container with id 038f171d600d493add18d92fee2bc0ed4fa96c59c9dcd002a109d1aa872acaf8 Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.822624 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" event={"ID":"d0c5a2fb-94ab-4f05-acba-73a19cb2e798","Type":"ContainerStarted","Data":"98dd4f093addc37c9d96e9499398ac986e240c778f467705a65fba5869279dee"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.836936 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-s6kbn" podStartSLOduration=124.836921666 podStartE2EDuration="2m4.836921666s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:37.8358214 +0000 UTC m=+155.024052175" watchObservedRunningTime="2025-12-04 06:13:37.836921666 +0000 UTC m=+155.025152441" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.848588 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.848943 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.348927774 +0000 UTC m=+155.537158559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.875051 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" event={"ID":"77c35356-f854-4d2d-a4d3-44c107d6f3bd","Type":"ContainerStarted","Data":"89d8254723bf51f8925486d15d6a1fd93eedd80a120a0d25d75210ccaf9e111d"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.903363 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" event={"ID":"452e5252-d838-4a36-96f4-f75ca800919f","Type":"ContainerStarted","Data":"9a00766081e01a130a42a70c7b6247d910ace46e7522310baca523be3d1d4d41"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.923754 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" event={"ID":"7736df92-b993-435a-ad74-4c2edd644ca3","Type":"ContainerStarted","Data":"7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.923806 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" event={"ID":"7736df92-b993-435a-ad74-4c2edd644ca3","Type":"ContainerStarted","Data":"587658f83a4372449257a5d4318019a90d96e2154c4dcaa48742344d95d3fd87"} Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.925276 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.937867 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-crn6w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.937919 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.949848 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:37 crc kubenswrapper[4685]: E1204 06:13:37.950969 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.450952394 +0000 UTC m=+155.639183169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:37 crc kubenswrapper[4685]: I1204 06:13:37.976865 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-qxx2n" event={"ID":"55269b01-b42d-4d38-8d36-2414f54adbe8","Type":"ContainerStarted","Data":"431ac77ab2db919c5e929c4ae2aa2c519ae6aaaae0e505d148d55f6113ce34ee"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.009264 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" event={"ID":"25743472-93a4-4d4e-bf54-1d5cef4025a2","Type":"ContainerStarted","Data":"b6a0ea9ea10f10becae2a56e03586850fc00c66c39dd524fdc18f143c155e4ee"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.011367 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" podStartSLOduration=124.011352816 podStartE2EDuration="2m4.011352816s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.009312389 +0000 UTC m=+155.197543164" watchObservedRunningTime="2025-12-04 06:13:38.011352816 +0000 UTC m=+155.199583591" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.037524 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" event={"ID":"f53d1c26-cef7-4963-aebc-bb7d8489f070","Type":"ContainerStarted","Data":"befd21522efb3bede32231652aa8cea47e199ba593f94f799af2942bdbf5f723"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.045821 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" event={"ID":"3e908422-dc37-44e1-abf1-d617bb1b7832","Type":"ContainerStarted","Data":"ddb6c2b747b05493af829d643b400a70572b4ff042c8c3341341123713b24315"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.048651 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-g9csq" event={"ID":"2e3c86c7-f36b-430d-9950-9542224eeee9","Type":"ContainerStarted","Data":"329c1cbd2af1925da34231b77d5aa1772dcff51aa6e709bf65344bb24dd47267"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.049082 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.050696 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.050964 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" event={"ID":"a55f07a2-0741-486a-8e5a-63c1b823e443","Type":"ContainerStarted","Data":"56ac5a7a1bc8149d6588d86273097ba950acb6d10f644119fa64afb6053db67a"} Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.052257 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.552239911 +0000 UTC m=+155.740470676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.060935 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" event={"ID":"bc81c705-faab-44b6-8da9-96bee4bd043a","Type":"ContainerStarted","Data":"ff06344560280d71471983d9204b61478e8388738c556b19cf347183a6485ed9"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.065306 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" event={"ID":"007053aa-6733-47ec-aeb6-7b6753d58c15","Type":"ContainerStarted","Data":"f32f9e4b055b2ad9e44d2c6c31a54e0973e81150fec55e65593ee130c390f205"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.081946 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" event={"ID":"9f122542-8cb9-494b-bf1b-fcae0350553a","Type":"ContainerStarted","Data":"f4ef29ff1b730594e13b16a4361ed42807b9bd02754e5e62e009a108d8798669"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.110829 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.115648 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.137054 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" event={"ID":"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2","Type":"ContainerStarted","Data":"b33f65c6690adc2a9b9dcc6343b36be713276affa28629cda095ab3309c2d5eb"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.152387 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" event={"ID":"f2e25031-7d34-4680-8bad-592c3ecb8764","Type":"ContainerStarted","Data":"656dac9e8c0d8b8a0a11c694983cf189ced6fada4c7e1b5ac26bde7331133435"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.153558 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" event={"ID":"7b2a9d09-996b-4659-8f4f-8583055b0894","Type":"ContainerStarted","Data":"2b49ea9e22aab2a5ea4926d1e414a5b181096b10a0491892dcbe106261c5ece8"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.153586 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" event={"ID":"7b2a9d09-996b-4659-8f4f-8583055b0894","Type":"ContainerStarted","Data":"a06b1ca419282b3cacabf4c61e8437bb0cabb43f5fd6fd719e894aac879f3278"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.155931 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.156121 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.656097183 +0000 UTC m=+155.844327958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.156235 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.159762 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.659747304 +0000 UTC m=+155.847978069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.165508 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-qxx2n" podStartSLOduration=125.165492504 podStartE2EDuration="2m5.165492504s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.068717247 +0000 UTC m=+155.256948022" watchObservedRunningTime="2025-12-04 06:13:38.165492504 +0000 UTC m=+155.353723279" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.167431 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" event={"ID":"6dd8d948-2f17-4ef8-8b31-572f08420bb2","Type":"ContainerStarted","Data":"a878c149edd1c072fbb8e0cc5b35eb1ed686301932ecbb402d12ec2f13833d0c"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.168085 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.184301 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" event={"ID":"3334bba6-dbe5-4be6-909e-95cd1b26d6ce","Type":"ContainerStarted","Data":"29b258193a723629f165c90845236430fe7861082068060e0f17f1932bc85c0d"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.184344 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" event={"ID":"3334bba6-dbe5-4be6-909e-95cd1b26d6ce","Type":"ContainerStarted","Data":"3a89545e64536dd3f9594eb50843ae7c93ab35dbf03afe5d38615ba879ce40d3"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.260105 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.273343 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.773320897 +0000 UTC m=+155.961551672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.282485 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vkf95" event={"ID":"dbd5e94d-05ab-4ac0-9846-e700839cce2a","Type":"ContainerStarted","Data":"42b8386a14f57618f724bd090851823eb32828045d8ea3bc1e40de803f55125b"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.284877 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" event={"ID":"5d290da7-657f-42e8-873f-74879e993568","Type":"ContainerStarted","Data":"a0936d9abd46fdb1df9015993154c6029cba39975e6a800ed2f35d1c188c77c1"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.286283 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" event={"ID":"9da20a29-ca25-4f24-97ef-d081cf757bac","Type":"ContainerStarted","Data":"21f085754ef263deccc53da5b301fa1e1cfd0d751f8828efbc1a81a6a2000354"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.302621 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" event={"ID":"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439","Type":"ContainerStarted","Data":"0b1e6e99412866a8c70232ca4999097c5d8fdc8d094536c3e91845bf3cd3c07d"} Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.318176 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2wd6d" podStartSLOduration=124.318159043 podStartE2EDuration="2m4.318159043s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.165871317 +0000 UTC m=+155.354102102" watchObservedRunningTime="2025-12-04 06:13:38.318159043 +0000 UTC m=+155.506389818" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.334282 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6zcm5" podStartSLOduration=125.334248076 podStartE2EDuration="2m5.334248076s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.319687184 +0000 UTC m=+155.507917989" watchObservedRunningTime="2025-12-04 06:13:38.334248076 +0000 UTC m=+155.522478861" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.347757 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-s6zxb" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.371043 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.371320 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.871305344 +0000 UTC m=+156.059536119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.414047 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9hw7" podStartSLOduration=124.41402561 podStartE2EDuration="2m4.41402561s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.398756664 +0000 UTC m=+155.586987439" watchObservedRunningTime="2025-12-04 06:13:38.41402561 +0000 UTC m=+155.602256385" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.469710 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-g9csq" podStartSLOduration=8.469692215 podStartE2EDuration="8.469692215s" podCreationTimestamp="2025-12-04 06:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.468537897 +0000 UTC m=+155.656768692" watchObservedRunningTime="2025-12-04 06:13:38.469692215 +0000 UTC m=+155.657922990" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.484476 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.484877 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:38.984857338 +0000 UTC m=+156.173088113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.535684 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7l82d" podStartSLOduration=125.535659641 podStartE2EDuration="2m5.535659641s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.52898961 +0000 UTC m=+155.717220385" watchObservedRunningTime="2025-12-04 06:13:38.535659641 +0000 UTC m=+155.723890426" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.592600 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.593007 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.092992411 +0000 UTC m=+156.281223186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.593989 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" podStartSLOduration=125.593977424 podStartE2EDuration="2m5.593977424s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.559744679 +0000 UTC m=+155.747975464" watchObservedRunningTime="2025-12-04 06:13:38.593977424 +0000 UTC m=+155.782208199" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.594940 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" podStartSLOduration=125.594933675 podStartE2EDuration="2m5.594933675s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.593083224 +0000 UTC m=+155.781313999" watchObservedRunningTime="2025-12-04 06:13:38.594933675 +0000 UTC m=+155.783164450" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.616171 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-q4sb7" podStartSLOduration=124.616154549 podStartE2EDuration="2m4.616154549s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:38.615800037 +0000 UTC m=+155.804030812" watchObservedRunningTime="2025-12-04 06:13:38.616154549 +0000 UTC m=+155.804385324" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.693728 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.694114 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.194098992 +0000 UTC m=+156.382329767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.734504 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.775134 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:38 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:38 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:38 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.775948 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.795275 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.795659 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.295645397 +0000 UTC m=+156.483876172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.896552 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.896902 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.396884312 +0000 UTC m=+156.585115087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.897156 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.897453 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.39744518 +0000 UTC m=+156.585675955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:38 crc kubenswrapper[4685]: I1204 06:13:38.997622 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:38 crc kubenswrapper[4685]: E1204 06:13:38.998062 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.498034643 +0000 UTC m=+156.686265428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.099657 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.099919 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.599903939 +0000 UTC m=+156.788134714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.200941 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.201104 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.701088083 +0000 UTC m=+156.889318858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.201346 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.201664 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.701655032 +0000 UTC m=+156.889885807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.203969 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-24w48" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.303318 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.303833 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.803812496 +0000 UTC m=+156.992043271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.304145 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.304532 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.80451053 +0000 UTC m=+156.992741305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.351219 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" event={"ID":"bc81c705-faab-44b6-8da9-96bee4bd043a","Type":"ContainerStarted","Data":"774b6a73a76a8a570017dfc9a3a611265d574edf851a53886a8fa38106e5e21a"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.358070 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" event={"ID":"9f122542-8cb9-494b-bf1b-fcae0350553a","Type":"ContainerStarted","Data":"2cd7564044ea207d0e027deec6773ec71902e634ed0f118127dfdfa34688575c"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.358114 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" event={"ID":"9f122542-8cb9-494b-bf1b-fcae0350553a","Type":"ContainerStarted","Data":"fae0cd08046dfeb697f0f953d820d0ef202916e83d62b66c4f572de72e284d32"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.358499 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.368723 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" event={"ID":"6b582cd3-5a8b-45ff-a6a7-05944059ae75","Type":"ContainerStarted","Data":"6628d20949c7882b4bee8de1d418457948f872f74bfec99451e82169faae8966"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.374746 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vkf95" event={"ID":"dbd5e94d-05ab-4ac0-9846-e700839cce2a","Type":"ContainerStarted","Data":"b6bbe8a9e4263c510a99797984b638365b532038fb778963663b17657d61620d"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.375370 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.377127 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" event={"ID":"77c35356-f854-4d2d-a4d3-44c107d6f3bd","Type":"ContainerStarted","Data":"085019bb8ea64d83e9a3650ba0bb802fc078132bf303d4481a00cd200bd228a9"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.384275 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-vkf95 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.384324 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vkf95" podUID="dbd5e94d-05ab-4ac0-9846-e700839cce2a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.384574 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qz8gl" event={"ID":"e8973864-fd68-4a8b-a298-e298f515a087","Type":"ContainerStarted","Data":"c6d384df8345308ab135c32b26e2679f3721931d0a6155490d51e8b35c150be8"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.386021 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" podStartSLOduration=125.38600227 podStartE2EDuration="2m5.38600227s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.385318307 +0000 UTC m=+156.573549082" watchObservedRunningTime="2025-12-04 06:13:39.38600227 +0000 UTC m=+156.574233045" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.405664 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.406774 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:39.906752057 +0000 UTC m=+157.094982832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.431529 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-g9csq" event={"ID":"2e3c86c7-f36b-430d-9950-9542224eeee9","Type":"ContainerStarted","Data":"b6bce2488b8027052dcba2cf1891e05cda6b0e2b791147516375227527e2866d"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.442656 4685 patch_prober.go:28] interesting pod/apiserver-76f77b778f-fz7dx container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]log ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]etcd ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/max-in-flight-filter ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 04 06:13:39 crc kubenswrapper[4685]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 04 06:13:39 crc kubenswrapper[4685]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/project.openshift.io-projectcache ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/openshift.io-startinformers ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 04 06:13:39 crc kubenswrapper[4685]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 04 06:13:39 crc kubenswrapper[4685]: livez check failed Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.442708 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" podUID="bf8ca363-9882-4e15-bfe7-5b59686e88ea" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.444169 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" event={"ID":"8e293db2-ec8b-454a-8e09-72aea57b2c97","Type":"ContainerStarted","Data":"2bb7bce2663147c17b9aee21b09d3b19e742e1896a4824157753e7d5f8a1fe95"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.461636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" event={"ID":"a55f07a2-0741-486a-8e5a-63c1b823e443","Type":"ContainerStarted","Data":"e0fc5f84b4785995725a53087ceaf3996f06f0fb3cfafe02f22b7e145c45f907"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.467048 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p5tm9" podStartSLOduration=126.467032136 podStartE2EDuration="2m6.467032136s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.424765445 +0000 UTC m=+156.612996220" watchObservedRunningTime="2025-12-04 06:13:39.467032136 +0000 UTC m=+156.655262911" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.470351 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" event={"ID":"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2","Type":"ContainerStarted","Data":"5f17491754b25d31331acbd6a1e6f41252fa9873da804e7619b10a8295a10013"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.481481 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" event={"ID":"452e5252-d838-4a36-96f4-f75ca800919f","Type":"ContainerStarted","Data":"2c1052040a256d04bc02ba6ff49ac8f478834184864c4e946b7a45c52cedcafc"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.501830 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.502344 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vkf95" podStartSLOduration=126.502322325 podStartE2EDuration="2m6.502322325s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.502089987 +0000 UTC m=+156.690320762" watchObservedRunningTime="2025-12-04 06:13:39.502322325 +0000 UTC m=+156.690553110" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.502515 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qz8gl" podStartSLOduration=8.502505701 podStartE2EDuration="8.502505701s" podCreationTimestamp="2025-12-04 06:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.467920635 +0000 UTC m=+156.656151410" watchObservedRunningTime="2025-12-04 06:13:39.502505701 +0000 UTC m=+156.690736506" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.507852 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.508935 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.008921104 +0000 UTC m=+157.197151879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.511309 4685 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-px7gh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.511359 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" podUID="452e5252-d838-4a36-96f4-f75ca800919f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.512190 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" event={"ID":"3334bba6-dbe5-4be6-909e-95cd1b26d6ce","Type":"ContainerStarted","Data":"092610e11be4dfc6e25c385f11bb21c51e9f3e04eeac4a98334985e0a8ab6021"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.528957 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvcw7" podStartSLOduration=125.528935527 podStartE2EDuration="2m5.528935527s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.517574841 +0000 UTC m=+156.705805616" watchObservedRunningTime="2025-12-04 06:13:39.528935527 +0000 UTC m=+156.717166302" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.537216 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" event={"ID":"f2e25031-7d34-4680-8bad-592c3ecb8764","Type":"ContainerStarted","Data":"3687a7d33038a093f33869ee90b8ec1fa22c9d36307f421b90369042eae53e57"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.538439 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tgrh2" podStartSLOduration=125.538422161 podStartE2EDuration="2m5.538422161s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.537786711 +0000 UTC m=+156.726017486" watchObservedRunningTime="2025-12-04 06:13:39.538422161 +0000 UTC m=+156.726652936" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.542945 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" event={"ID":"9da20a29-ca25-4f24-97ef-d081cf757bac","Type":"ContainerStarted","Data":"a635ec87c12129adcd8a1f64effe8820e196ceb156ec477dc1b1aefef7848648"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.546150 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.552837 4685 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gt6kp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.552884 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" podUID="9da20a29-ca25-4f24-97ef-d081cf757bac" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.555808 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" event={"ID":"5d290da7-657f-42e8-873f-74879e993568","Type":"ContainerStarted","Data":"bc5f049c5ff13bc138af4521e54449a0d17f50f80c531540589620d2ec2a8a71"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.579579 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4d924d97f115425dcd319c01fc82f727f2867d9eafec603f7626bac0c12b188a"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.579620 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"038f171d600d493add18d92fee2bc0ed4fa96c59c9dcd002a109d1aa872acaf8"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.580128 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.601742 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d251a2cc66ed5632d56815caa037a4f6b0767d85962eae5814af4c01d5aec148"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.601785 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"735a58d52edf84d0ed6635f8e8bac74d94e9827488cf7cfa5210d5a92d35607d"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.608975 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.610178 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.110162649 +0000 UTC m=+157.298393424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.619705 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pwwjm" podStartSLOduration=126.619688374 podStartE2EDuration="2m6.619688374s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.619485027 +0000 UTC m=+156.807715802" watchObservedRunningTime="2025-12-04 06:13:39.619688374 +0000 UTC m=+156.807919139" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.620028 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" podStartSLOduration=125.620024075 podStartE2EDuration="2m5.620024075s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.574191997 +0000 UTC m=+156.762422772" watchObservedRunningTime="2025-12-04 06:13:39.620024075 +0000 UTC m=+156.808254850" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.621841 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" event={"ID":"530a634c-fa68-456c-afbc-45f5a15151f1","Type":"ContainerStarted","Data":"5f4a05d2c5fe2baa9ac19005e345fb4714d9df41f629950d68c84124ce548f59"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.634754 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" event={"ID":"63fa9dc9-ddb8-4731-ad67-cec8fb8a7439","Type":"ContainerStarted","Data":"d0781e7846913dc05f374d8983872c47c96a50acaadd8962c0dd286a0a66e3a8"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.635953 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.644526 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" podStartSLOduration=125.644508516 podStartE2EDuration="2m5.644508516s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.644493966 +0000 UTC m=+156.832724751" watchObservedRunningTime="2025-12-04 06:13:39.644508516 +0000 UTC m=+156.832739291" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.654289 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"36d8a6a295e17c395175a2c4e834252b69ca8d116faf3b5a05273e12b1b61486"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.669509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" event={"ID":"89e148da-1707-4caa-9e28-7976df0303a5","Type":"ContainerStarted","Data":"ecf0025d1b40f24ebc67c6fde53f5804bec18e6413ca13f40ebd6b6d90c63aa5"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.687431 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" event={"ID":"d0c5a2fb-94ab-4f05-acba-73a19cb2e798","Type":"ContainerStarted","Data":"62939e6ef508b2d035a9772465be519683baec23f2ecbf7e0f2eae979548cc3d"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.687474 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" event={"ID":"d0c5a2fb-94ab-4f05-acba-73a19cb2e798","Type":"ContainerStarted","Data":"2d7528b2b8db1627f5851aaf8ea2738799da05d268b27d728de02d7e5e27c37d"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.697361 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pwfqr" podStartSLOduration=125.697345918 podStartE2EDuration="2m5.697345918s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.668396388 +0000 UTC m=+156.856627163" watchObservedRunningTime="2025-12-04 06:13:39.697345918 +0000 UTC m=+156.885576693" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.706185 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" event={"ID":"02ef6855-1de7-411b-9dca-e8a777d0f461","Type":"ContainerStarted","Data":"007fee82e3b739378c77a1f56cbcc1b1ff334ae04f5e5e1aa44e8a034acc8f3a"} Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.708143 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-crn6w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.708191 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.710178 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.711702 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.211686733 +0000 UTC m=+157.399917558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.740095 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:39 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:39 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:39 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.740143 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.769431 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" podStartSLOduration=125.769388715 podStartE2EDuration="2m5.769388715s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.768519626 +0000 UTC m=+156.956750421" watchObservedRunningTime="2025-12-04 06:13:39.769388715 +0000 UTC m=+156.957619490" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.800528 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" podStartSLOduration=125.800513457 podStartE2EDuration="2m5.800513457s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.799842734 +0000 UTC m=+156.988073509" watchObservedRunningTime="2025-12-04 06:13:39.800513457 +0000 UTC m=+156.988744232" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.810740 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.812537 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.312523944 +0000 UTC m=+157.500754719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.831458 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" podStartSLOduration=125.831439552 podStartE2EDuration="2m5.831439552s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.830860172 +0000 UTC m=+157.019090947" watchObservedRunningTime="2025-12-04 06:13:39.831439552 +0000 UTC m=+157.019670327" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.854640 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lkbzh" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.862458 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cvsf2" podStartSLOduration=126.862438629 podStartE2EDuration="2m6.862438629s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.859640396 +0000 UTC m=+157.047871171" watchObservedRunningTime="2025-12-04 06:13:39.862438629 +0000 UTC m=+157.050669404" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.887264 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6j5ws" podStartSLOduration=126.887249401 podStartE2EDuration="2m6.887249401s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:39.88630554 +0000 UTC m=+157.074536315" watchObservedRunningTime="2025-12-04 06:13:39.887249401 +0000 UTC m=+157.075480176" Dec 04 06:13:39 crc kubenswrapper[4685]: I1204 06:13:39.913271 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:39 crc kubenswrapper[4685]: E1204 06:13:39.913607 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.413594344 +0000 UTC m=+157.601825119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.014598 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.014835 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.514802778 +0000 UTC m=+157.703033553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.014999 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.015319 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.515306575 +0000 UTC m=+157.703537350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.116780 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.116962 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.616937613 +0000 UTC m=+157.805168388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.117403 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.117692 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.617679827 +0000 UTC m=+157.805910602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.218695 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.219088 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.719059686 +0000 UTC m=+157.907290461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.219306 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.219664 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.719646816 +0000 UTC m=+157.907877601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.320320 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.320531 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.820505548 +0000 UTC m=+158.008736323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.320633 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.320961 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.820948743 +0000 UTC m=+158.009179518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.421599 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.421800 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.921770884 +0000 UTC m=+158.110001689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.421872 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.422129 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:40.922121396 +0000 UTC m=+158.110352171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.523448 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.523660 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.02363582 +0000 UTC m=+158.211866595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.523742 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.524044 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.024030713 +0000 UTC m=+158.212261488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.559149 4685 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.624794 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.624979 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.124951987 +0000 UTC m=+158.313182762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.625578 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.625874 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.125864628 +0000 UTC m=+158.314095403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.717688 4685 generic.go:334] "Generic (PLEG): container finished" podID="bc81c705-faab-44b6-8da9-96bee4bd043a" containerID="774b6a73a76a8a570017dfc9a3a611265d574edf851a53886a8fa38106e5e21a" exitCode=0 Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.717755 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" event={"ID":"bc81c705-faab-44b6-8da9-96bee4bd043a","Type":"ContainerDied","Data":"774b6a73a76a8a570017dfc9a3a611265d574edf851a53886a8fa38106e5e21a"} Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.722170 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" event={"ID":"c36c0f14-f994-4969-bb8e-f5d6e7be4bf2","Type":"ContainerStarted","Data":"4a157a4b7695ce8ce2311db6ca96ed607d53838a6e9cd9cabddf71cef012d767"} Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.724297 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pcxg4" event={"ID":"530a634c-fa68-456c-afbc-45f5a15151f1","Type":"ContainerStarted","Data":"10905a0c70e40dbe19106327630cfe5e5ffffc5ccafc71c71c057bba6a85fccf"} Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.726363 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.726492 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.226476152 +0000 UTC m=+158.414706927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.726544 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.726862 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.226850354 +0000 UTC m=+158.415081129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.726895 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" event={"ID":"f2e25031-7d34-4680-8bad-592c3ecb8764","Type":"ContainerStarted","Data":"3ad3e1cc3535857757e060ef9e8029ef57c4a5733a39a270be4c6738512936c3"} Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.726915 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" event={"ID":"f2e25031-7d34-4680-8bad-592c3ecb8764","Type":"ContainerStarted","Data":"b99cbfedbc0cccbe444db73199d35c41ce3c9317e9ed1cfaecf931aef64290d7"} Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.730880 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" event={"ID":"77c35356-f854-4d2d-a4d3-44c107d6f3bd","Type":"ContainerStarted","Data":"830a4bbb98be79e3e117cae5fd946c41503fb35318f81b27995d78e9ecec5af5"} Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.731770 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-vkf95 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.731803 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vkf95" podUID="dbd5e94d-05ab-4ac0-9846-e700839cce2a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.742222 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:40 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:40 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:40 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.742298 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.747699 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-px7gh" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.757321 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.761628 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt6kp" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.788259 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8ndqp" podStartSLOduration=126.788244509 podStartE2EDuration="2m6.788244509s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:40.78644659 +0000 UTC m=+157.974677365" watchObservedRunningTime="2025-12-04 06:13:40.788244509 +0000 UTC m=+157.976475284" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.805897 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tgb7c" podStartSLOduration=126.805881924 podStartE2EDuration="2m6.805881924s" podCreationTimestamp="2025-12-04 06:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:40.804128375 +0000 UTC m=+157.992359150" watchObservedRunningTime="2025-12-04 06:13:40.805881924 +0000 UTC m=+157.994112699" Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.830938 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.832529 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.332508555 +0000 UTC m=+158.520739330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:40 crc kubenswrapper[4685]: I1204 06:13:40.935667 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:40 crc kubenswrapper[4685]: E1204 06:13:40.936054 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.436038656 +0000 UTC m=+158.624269431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.036571 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:41 crc kubenswrapper[4685]: E1204 06:13:41.036713 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.536688762 +0000 UTC m=+158.724919537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.036869 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:41 crc kubenswrapper[4685]: E1204 06:13:41.037208 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.537199599 +0000 UTC m=+158.725430374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.139111 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:41 crc kubenswrapper[4685]: E1204 06:13:41.139393 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.639377645 +0000 UTC m=+158.827608420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.240995 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:41 crc kubenswrapper[4685]: E1204 06:13:41.241494 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 06:13:41.741473548 +0000 UTC m=+158.929704323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jkj6h" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.275499 4685 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T06:13:40.559182188Z","Handler":null,"Name":""} Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.281511 4685 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.281536 4685 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.318205 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xcp78"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.319488 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.321898 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.329129 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xcp78"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.344146 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.352014 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.444998 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz4b6\" (UniqueName: \"kubernetes.io/projected/b2596408-3b90-481e-ad56-d5560edf0a96-kube-api-access-zz4b6\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.445128 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.445159 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-utilities\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.445231 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-catalog-content\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.448357 4685 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.448425 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.493432 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jkj6h\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.524010 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2hrbf"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.525252 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.534228 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2hrbf"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.537084 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.546471 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-utilities\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.546545 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-catalog-content\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.546603 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz4b6\" (UniqueName: \"kubernetes.io/projected/b2596408-3b90-481e-ad56-d5560edf0a96-kube-api-access-zz4b6\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.547124 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-utilities\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.547128 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-catalog-content\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.575177 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz4b6\" (UniqueName: \"kubernetes.io/projected/b2596408-3b90-481e-ad56-d5560edf0a96-kube-api-access-zz4b6\") pod \"certified-operators-xcp78\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.635075 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.647505 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-utilities\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.647619 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr6rl\" (UniqueName: \"kubernetes.io/projected/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-kube-api-access-hr6rl\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.647706 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-catalog-content\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.725656 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-47bq7"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.726716 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.729456 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-47bq7"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.740688 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:41 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:41 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:41 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.740742 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.741342 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.753497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-utilities\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.753555 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr6rl\" (UniqueName: \"kubernetes.io/projected/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-kube-api-access-hr6rl\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.753584 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-catalog-content\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.754376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-utilities\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.754387 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-catalog-content\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.811216 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" event={"ID":"f2e25031-7d34-4680-8bad-592c3ecb8764","Type":"ContainerStarted","Data":"0b3f523937277f75c0e1bbff34f64dcf1d98ff6e70e4fa0d986a15d1dc2eab55"} Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.817324 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr6rl\" (UniqueName: \"kubernetes.io/projected/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-kube-api-access-hr6rl\") pod \"community-operators-2hrbf\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.822593 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-vkf95 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.822658 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vkf95" podUID="dbd5e94d-05ab-4ac0-9846-e700839cce2a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.846586 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.856986 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-utilities\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.857172 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-kube-api-access-vnjlp\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.857230 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-catalog-content\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.857503 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jjm5p" podStartSLOduration=11.857476112 podStartE2EDuration="11.857476112s" podCreationTimestamp="2025-12-04 06:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:41.847586894 +0000 UTC m=+159.035817679" watchObservedRunningTime="2025-12-04 06:13:41.857476112 +0000 UTC m=+159.045706897" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.919866 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-872w4"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.920900 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.927344 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-872w4"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960051 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-utilities\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960090 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdn2n\" (UniqueName: \"kubernetes.io/projected/4121c660-39c8-4dd8-8a26-f40f37fb16bd-kube-api-access-kdn2n\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960150 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-kube-api-access-vnjlp\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960171 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-catalog-content\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960193 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-utilities\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960214 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-catalog-content\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.960601 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-catalog-content\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.961536 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-utilities\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.964195 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xcp78"] Dec 04 06:13:41 crc kubenswrapper[4685]: I1204 06:13:41.978461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-kube-api-access-vnjlp\") pod \"certified-operators-47bq7\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.061168 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdn2n\" (UniqueName: \"kubernetes.io/projected/4121c660-39c8-4dd8-8a26-f40f37fb16bd-kube-api-access-kdn2n\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.061244 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-catalog-content\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.061265 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-utilities\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.061902 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-utilities\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.061941 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-catalog-content\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.075340 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.079537 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdn2n\" (UniqueName: \"kubernetes.io/projected/4121c660-39c8-4dd8-8a26-f40f37fb16bd-kube-api-access-kdn2n\") pod \"community-operators-872w4\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.123247 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.162303 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc81c705-faab-44b6-8da9-96bee4bd043a-config-volume\") pod \"bc81c705-faab-44b6-8da9-96bee4bd043a\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.162346 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhkzx\" (UniqueName: \"kubernetes.io/projected/bc81c705-faab-44b6-8da9-96bee4bd043a-kube-api-access-zhkzx\") pod \"bc81c705-faab-44b6-8da9-96bee4bd043a\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.162433 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc81c705-faab-44b6-8da9-96bee4bd043a-secret-volume\") pod \"bc81c705-faab-44b6-8da9-96bee4bd043a\" (UID: \"bc81c705-faab-44b6-8da9-96bee4bd043a\") " Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.163776 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc81c705-faab-44b6-8da9-96bee4bd043a-config-volume" (OuterVolumeSpecName: "config-volume") pod "bc81c705-faab-44b6-8da9-96bee4bd043a" (UID: "bc81c705-faab-44b6-8da9-96bee4bd043a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.168360 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc81c705-faab-44b6-8da9-96bee4bd043a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bc81c705-faab-44b6-8da9-96bee4bd043a" (UID: "bc81c705-faab-44b6-8da9-96bee4bd043a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.168564 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc81c705-faab-44b6-8da9-96bee4bd043a-kube-api-access-zhkzx" (OuterVolumeSpecName: "kube-api-access-zhkzx") pod "bc81c705-faab-44b6-8da9-96bee4bd043a" (UID: "bc81c705-faab-44b6-8da9-96bee4bd043a"). InnerVolumeSpecName "kube-api-access-zhkzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.169904 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2hrbf"] Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.248197 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.266482 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc81c705-faab-44b6-8da9-96bee4bd043a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.266509 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc81c705-faab-44b6-8da9-96bee4bd043a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.266520 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhkzx\" (UniqueName: \"kubernetes.io/projected/bc81c705-faab-44b6-8da9-96bee4bd043a-kube-api-access-zhkzx\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.298754 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-47bq7"] Dec 04 06:13:42 crc kubenswrapper[4685]: W1204 06:13:42.322557 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod156c8e99_d0bf_4c3f_bf15_cf0a50113b03.slice/crio-60d82f3fd7c19e9a17a7892efd208f25bcb0b687db5d81ba70d7d162d9637d11 WatchSource:0}: Error finding container 60d82f3fd7c19e9a17a7892efd208f25bcb0b687db5d81ba70d7d162d9637d11: Status 404 returned error can't find the container with id 60d82f3fd7c19e9a17a7892efd208f25bcb0b687db5d81ba70d7d162d9637d11 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.330217 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jkj6h"] Dec 04 06:13:42 crc kubenswrapper[4685]: W1204 06:13:42.443691 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfebae4b_a7c1_47a0_9efa_4c7127df9a13.slice/crio-2cca135bf3eda5692fa832aa05e27278bd2a5a5554ebc46110aeba0ca0c15cf6 WatchSource:0}: Error finding container 2cca135bf3eda5692fa832aa05e27278bd2a5a5554ebc46110aeba0ca0c15cf6: Status 404 returned error can't find the container with id 2cca135bf3eda5692fa832aa05e27278bd2a5a5554ebc46110aeba0ca0c15cf6 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.497203 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-872w4"] Dec 04 06:13:42 crc kubenswrapper[4685]: W1204 06:13:42.509266 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4121c660_39c8_4dd8_8a26_f40f37fb16bd.slice/crio-dc38eda2f8f62624c9a39a370df183068c52a1bfecf53858170da901281d3435 WatchSource:0}: Error finding container dc38eda2f8f62624c9a39a370df183068c52a1bfecf53858170da901281d3435: Status 404 returned error can't find the container with id dc38eda2f8f62624c9a39a370df183068c52a1bfecf53858170da901281d3435 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.736767 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:42 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:42 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:42 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.736818 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.797575 4685 generic.go:334] "Generic (PLEG): container finished" podID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerID="b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d" exitCode=0 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.797662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hrbf" event={"ID":"b1a21c41-db1c-40c9-82d1-2fb0f33f4625","Type":"ContainerDied","Data":"b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.797695 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hrbf" event={"ID":"b1a21c41-db1c-40c9-82d1-2fb0f33f4625","Type":"ContainerStarted","Data":"aec6c76df4f77c213ea64af7070859b4ccfeddf45a8a65361e319c6da6f63d65"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.799187 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.799975 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" event={"ID":"bc81c705-faab-44b6-8da9-96bee4bd043a","Type":"ContainerDied","Data":"ff06344560280d71471983d9204b61478e8388738c556b19cf347183a6485ed9"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.800000 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.800006 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff06344560280d71471983d9204b61478e8388738c556b19cf347183a6485ed9" Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.801350 4685 generic.go:334] "Generic (PLEG): container finished" podID="b2596408-3b90-481e-ad56-d5560edf0a96" containerID="f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196" exitCode=0 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.801401 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcp78" event={"ID":"b2596408-3b90-481e-ad56-d5560edf0a96","Type":"ContainerDied","Data":"f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.801439 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcp78" event={"ID":"b2596408-3b90-481e-ad56-d5560edf0a96","Type":"ContainerStarted","Data":"d3314796cffe16cc933e05695536cae8b78c7c50ba2b4ece75c0fa8f92fb483e"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.803742 4685 generic.go:334] "Generic (PLEG): container finished" podID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerID="ad3d1a57e8c04ec809d7b9e4342fbf280d59c273a8169686bc647474de001ea2" exitCode=0 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.803825 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-47bq7" event={"ID":"156c8e99-d0bf-4c3f-bf15-cf0a50113b03","Type":"ContainerDied","Data":"ad3d1a57e8c04ec809d7b9e4342fbf280d59c273a8169686bc647474de001ea2"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.803858 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-47bq7" event={"ID":"156c8e99-d0bf-4c3f-bf15-cf0a50113b03","Type":"ContainerStarted","Data":"60d82f3fd7c19e9a17a7892efd208f25bcb0b687db5d81ba70d7d162d9637d11"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.806431 4685 generic.go:334] "Generic (PLEG): container finished" podID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerID="ef0efcc27ddd3cffc17f82771ddc7b7e1d4f2bc5e3b1711327148d13b4f3ecfc" exitCode=0 Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.806497 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872w4" event={"ID":"4121c660-39c8-4dd8-8a26-f40f37fb16bd","Type":"ContainerDied","Data":"ef0efcc27ddd3cffc17f82771ddc7b7e1d4f2bc5e3b1711327148d13b4f3ecfc"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.806516 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872w4" event={"ID":"4121c660-39c8-4dd8-8a26-f40f37fb16bd","Type":"ContainerStarted","Data":"dc38eda2f8f62624c9a39a370df183068c52a1bfecf53858170da901281d3435"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.808176 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" event={"ID":"bfebae4b-a7c1-47a0-9efa-4c7127df9a13","Type":"ContainerStarted","Data":"a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.808215 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" event={"ID":"bfebae4b-a7c1-47a0-9efa-4c7127df9a13","Type":"ContainerStarted","Data":"2cca135bf3eda5692fa832aa05e27278bd2a5a5554ebc46110aeba0ca0c15cf6"} Dec 04 06:13:42 crc kubenswrapper[4685]: I1204 06:13:42.909125 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" podStartSLOduration=129.909111011 podStartE2EDuration="2m9.909111011s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:42.907007982 +0000 UTC m=+160.095238757" watchObservedRunningTime="2025-12-04 06:13:42.909111011 +0000 UTC m=+160.097341786" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.110978 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.115926 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-fz7dx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.119802 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.119848 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.144744 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.145378 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.316096 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6tt6d"] Dec 04 06:13:43 crc kubenswrapper[4685]: E1204 06:13:43.316294 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc81c705-faab-44b6-8da9-96bee4bd043a" containerName="collect-profiles" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.316305 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc81c705-faab-44b6-8da9-96bee4bd043a" containerName="collect-profiles" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.316420 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc81c705-faab-44b6-8da9-96bee4bd043a" containerName="collect-profiles" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.317190 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.321889 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.329452 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tt6d"] Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.380333 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-catalog-content\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.380464 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-utilities\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.380499 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpzzx\" (UniqueName: \"kubernetes.io/projected/ca2b22c7-e7ef-445f-a910-d068f1e17575-kube-api-access-fpzzx\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.481956 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-catalog-content\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.482609 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-catalog-content\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.482726 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-utilities\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.482764 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpzzx\" (UniqueName: \"kubernetes.io/projected/ca2b22c7-e7ef-445f-a910-d068f1e17575-kube-api-access-fpzzx\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.483548 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-utilities\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.513756 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpzzx\" (UniqueName: \"kubernetes.io/projected/ca2b22c7-e7ef-445f-a910-d068f1e17575-kube-api-access-fpzzx\") pod \"redhat-marketplace-6tt6d\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.635771 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.732535 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-frsvx"] Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.733979 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.734867 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.736864 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frsvx"] Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.745783 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:43 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:43 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:43 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.745841 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.787574 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vjks\" (UniqueName: \"kubernetes.io/projected/99e5eb96-f490-4718-8245-e7a0adee43a8-kube-api-access-4vjks\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.787721 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-utilities\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.787777 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-catalog-content\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.814835 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.814943 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.816839 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.817241 4685 patch_prober.go:28] interesting pod/console-f9d7485db-s6kbn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.817274 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-s6kbn" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.824099 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jmbn8" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.828914 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-vkf95 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.828952 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vkf95" podUID="dbd5e94d-05ab-4ac0-9846-e700839cce2a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.828986 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-vkf95 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.829045 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vkf95" podUID="dbd5e94d-05ab-4ac0-9846-e700839cce2a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.890308 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vjks\" (UniqueName: \"kubernetes.io/projected/99e5eb96-f490-4718-8245-e7a0adee43a8-kube-api-access-4vjks\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.890468 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-utilities\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.890531 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-catalog-content\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.893286 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-utilities\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.907773 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-catalog-content\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:43 crc kubenswrapper[4685]: I1204 06:13:43.916472 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vjks\" (UniqueName: \"kubernetes.io/projected/99e5eb96-f490-4718-8245-e7a0adee43a8-kube-api-access-4vjks\") pod \"redhat-marketplace-frsvx\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.061718 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.117754 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tt6d"] Dec 04 06:13:44 crc kubenswrapper[4685]: W1204 06:13:44.137466 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca2b22c7_e7ef_445f_a910_d068f1e17575.slice/crio-c31577c6525d1724d066c6b34dec76f775bc66911f4e26af918e3d8672598578 WatchSource:0}: Error finding container c31577c6525d1724d066c6b34dec76f775bc66911f4e26af918e3d8672598578: Status 404 returned error can't find the container with id c31577c6525d1724d066c6b34dec76f775bc66911f4e26af918e3d8672598578 Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.316723 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frsvx"] Dec 04 06:13:44 crc kubenswrapper[4685]: W1204 06:13:44.332857 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99e5eb96_f490_4718_8245_e7a0adee43a8.slice/crio-c43073d9090c5dfa98f8ae01913a87e8a332a1c284eca3d4374339311f9fdfac WatchSource:0}: Error finding container c43073d9090c5dfa98f8ae01913a87e8a332a1c284eca3d4374339311f9fdfac: Status 404 returned error can't find the container with id c43073d9090c5dfa98f8ae01913a87e8a332a1c284eca3d4374339311f9fdfac Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.520733 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-892qt"] Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.522007 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.524358 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.530325 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-892qt"] Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.604862 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.606159 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.610784 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.611009 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.618653 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpz72\" (UniqueName: \"kubernetes.io/projected/e70415f0-e1e4-4da7-a46a-ae3a01efa173-kube-api-access-xpz72\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.618784 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-utilities\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.618862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-catalog-content\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.623271 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.720697 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58a129c-f8ce-4b81-8293-30e5a31373fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.720764 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-utilities\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.720799 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-catalog-content\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.720823 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58a129c-f8ce-4b81-8293-30e5a31373fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.720873 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpz72\" (UniqueName: \"kubernetes.io/projected/e70415f0-e1e4-4da7-a46a-ae3a01efa173-kube-api-access-xpz72\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.721217 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-utilities\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.737587 4685 patch_prober.go:28] interesting pod/router-default-5444994796-qxx2n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 06:13:44 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Dec 04 06:13:44 crc kubenswrapper[4685]: [+]process-running ok Dec 04 06:13:44 crc kubenswrapper[4685]: healthz check failed Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.737628 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qxx2n" podUID="55269b01-b42d-4d38-8d36-2414f54adbe8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.752690 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-catalog-content\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.756889 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpz72\" (UniqueName: \"kubernetes.io/projected/e70415f0-e1e4-4da7-a46a-ae3a01efa173-kube-api-access-xpz72\") pod \"redhat-operators-892qt\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.821590 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58a129c-f8ce-4b81-8293-30e5a31373fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.821683 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58a129c-f8ce-4b81-8293-30e5a31373fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.821780 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58a129c-f8ce-4b81-8293-30e5a31373fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.822444 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frsvx" event={"ID":"99e5eb96-f490-4718-8245-e7a0adee43a8","Type":"ContainerStarted","Data":"c43073d9090c5dfa98f8ae01913a87e8a332a1c284eca3d4374339311f9fdfac"} Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.826422 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tt6d" event={"ID":"ca2b22c7-e7ef-445f-a910-d068f1e17575","Type":"ContainerStarted","Data":"c31577c6525d1724d066c6b34dec76f775bc66911f4e26af918e3d8672598578"} Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.858165 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58a129c-f8ce-4b81-8293-30e5a31373fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.858494 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.923828 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g9dg5"] Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.925087 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.930052 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:44 crc kubenswrapper[4685]: I1204 06:13:44.957831 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g9dg5"] Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.025370 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-utilities\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.025442 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bthx6\" (UniqueName: \"kubernetes.io/projected/fc7be107-549f-4c6e-86ea-0792e58cb2ab-kube-api-access-bthx6\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.025461 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-catalog-content\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.127507 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-utilities\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.127594 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bthx6\" (UniqueName: \"kubernetes.io/projected/fc7be107-549f-4c6e-86ea-0792e58cb2ab-kube-api-access-bthx6\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.127634 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-catalog-content\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.128514 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-catalog-content\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.128895 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-utilities\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.167975 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bthx6\" (UniqueName: \"kubernetes.io/projected/fc7be107-549f-4c6e-86ea-0792e58cb2ab-kube-api-access-bthx6\") pod \"redhat-operators-g9dg5\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.243747 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.287736 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.353026 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-892qt"] Dec 04 06:13:45 crc kubenswrapper[4685]: W1204 06:13:45.417040 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode70415f0_e1e4_4da7_a46a_ae3a01efa173.slice/crio-69196d35d40f446733c9567195b0bc8a4f157a42c240811999b54ed27635d3fc WatchSource:0}: Error finding container 69196d35d40f446733c9567195b0bc8a4f157a42c240811999b54ed27635d3fc: Status 404 returned error can't find the container with id 69196d35d40f446733c9567195b0bc8a4f157a42c240811999b54ed27635d3fc Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.739216 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.748142 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-qxx2n" Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.856942 4685 generic.go:334] "Generic (PLEG): container finished" podID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerID="b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a" exitCode=0 Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.857026 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerDied","Data":"b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a"} Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.857057 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerStarted","Data":"69196d35d40f446733c9567195b0bc8a4f157a42c240811999b54ed27635d3fc"} Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.867635 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b58a129c-f8ce-4b81-8293-30e5a31373fb","Type":"ContainerStarted","Data":"ea346bcbe26afabc3f4f92ae7641bc5e5d08b15ebdbf8531c70453770cc93232"} Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.874155 4685 generic.go:334] "Generic (PLEG): container finished" podID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerID="f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f" exitCode=0 Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.874235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frsvx" event={"ID":"99e5eb96-f490-4718-8245-e7a0adee43a8","Type":"ContainerDied","Data":"f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f"} Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.878197 4685 generic.go:334] "Generic (PLEG): container finished" podID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerID="2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688" exitCode=0 Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.878247 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tt6d" event={"ID":"ca2b22c7-e7ef-445f-a910-d068f1e17575","Type":"ContainerDied","Data":"2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688"} Dec 04 06:13:45 crc kubenswrapper[4685]: I1204 06:13:45.944692 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g9dg5"] Dec 04 06:13:45 crc kubenswrapper[4685]: W1204 06:13:45.966521 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc7be107_549f_4c6e_86ea_0792e58cb2ab.slice/crio-c0a03220f609875825cae8a56ada47d7ccbac3cce906ccaf3692f94b685b29c1 WatchSource:0}: Error finding container c0a03220f609875825cae8a56ada47d7ccbac3cce906ccaf3692f94b685b29c1: Status 404 returned error can't find the container with id c0a03220f609875825cae8a56ada47d7ccbac3cce906ccaf3692f94b685b29c1 Dec 04 06:13:46 crc kubenswrapper[4685]: I1204 06:13:46.178841 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-g9csq" Dec 04 06:13:46 crc kubenswrapper[4685]: I1204 06:13:46.907651 4685 generic.go:334] "Generic (PLEG): container finished" podID="b58a129c-f8ce-4b81-8293-30e5a31373fb" containerID="ac58558b148a84d76a9cfaf5df5f41d961f3afc413d0789e7101f69e566312c8" exitCode=0 Dec 04 06:13:46 crc kubenswrapper[4685]: I1204 06:13:46.907765 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b58a129c-f8ce-4b81-8293-30e5a31373fb","Type":"ContainerDied","Data":"ac58558b148a84d76a9cfaf5df5f41d961f3afc413d0789e7101f69e566312c8"} Dec 04 06:13:46 crc kubenswrapper[4685]: I1204 06:13:46.912403 4685 generic.go:334] "Generic (PLEG): container finished" podID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerID="3ceae89b6a00a83bb39cae0184d3a6ce8ef1a998652db68922e3ceb0bba44b49" exitCode=0 Dec 04 06:13:46 crc kubenswrapper[4685]: I1204 06:13:46.912448 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9dg5" event={"ID":"fc7be107-549f-4c6e-86ea-0792e58cb2ab","Type":"ContainerDied","Data":"3ceae89b6a00a83bb39cae0184d3a6ce8ef1a998652db68922e3ceb0bba44b49"} Dec 04 06:13:46 crc kubenswrapper[4685]: I1204 06:13:46.912462 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9dg5" event={"ID":"fc7be107-549f-4c6e-86ea-0792e58cb2ab","Type":"ContainerStarted","Data":"c0a03220f609875825cae8a56ada47d7ccbac3cce906ccaf3692f94b685b29c1"} Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.677263 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.678095 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.685349 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.685599 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.701600 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.799657 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.799738 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.901010 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.901071 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.901163 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:47 crc kubenswrapper[4685]: I1204 06:13:47.919929 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.000968 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.239100 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.275162 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.310211 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58a129c-f8ce-4b81-8293-30e5a31373fb-kube-api-access\") pod \"b58a129c-f8ce-4b81-8293-30e5a31373fb\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.310281 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58a129c-f8ce-4b81-8293-30e5a31373fb-kubelet-dir\") pod \"b58a129c-f8ce-4b81-8293-30e5a31373fb\" (UID: \"b58a129c-f8ce-4b81-8293-30e5a31373fb\") " Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.310863 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b58a129c-f8ce-4b81-8293-30e5a31373fb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b58a129c-f8ce-4b81-8293-30e5a31373fb" (UID: "b58a129c-f8ce-4b81-8293-30e5a31373fb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.316947 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b58a129c-f8ce-4b81-8293-30e5a31373fb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b58a129c-f8ce-4b81-8293-30e5a31373fb" (UID: "b58a129c-f8ce-4b81-8293-30e5a31373fb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.411855 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58a129c-f8ce-4b81-8293-30e5a31373fb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.411899 4685 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58a129c-f8ce-4b81-8293-30e5a31373fb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.941636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5ec886d-9e12-44c1-97c4-7ca9597174cc","Type":"ContainerStarted","Data":"925e1035d618a292e7701412aaae02ee9a91578113aeb368277c19f8c45fb133"} Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.949192 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b58a129c-f8ce-4b81-8293-30e5a31373fb","Type":"ContainerDied","Data":"ea346bcbe26afabc3f4f92ae7641bc5e5d08b15ebdbf8531c70453770cc93232"} Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.949229 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea346bcbe26afabc3f4f92ae7641bc5e5d08b15ebdbf8531c70453770cc93232" Dec 04 06:13:48 crc kubenswrapper[4685]: I1204 06:13:48.949321 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 06:13:49 crc kubenswrapper[4685]: I1204 06:13:49.964534 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5ec886d-9e12-44c1-97c4-7ca9597174cc","Type":"ContainerStarted","Data":"ae3489794b6b378cc85c08d11244d6cbd751684b4d15244e9a0c41ddd62ca17e"} Dec 04 06:13:49 crc kubenswrapper[4685]: I1204 06:13:49.979137 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.979123323 podStartE2EDuration="2.979123323s" podCreationTimestamp="2025-12-04 06:13:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:13:49.977695445 +0000 UTC m=+167.165926220" watchObservedRunningTime="2025-12-04 06:13:49.979123323 +0000 UTC m=+167.167354098" Dec 04 06:13:50 crc kubenswrapper[4685]: I1204 06:13:50.971018 4685 generic.go:334] "Generic (PLEG): container finished" podID="f5ec886d-9e12-44c1-97c4-7ca9597174cc" containerID="ae3489794b6b378cc85c08d11244d6cbd751684b4d15244e9a0c41ddd62ca17e" exitCode=0 Dec 04 06:13:50 crc kubenswrapper[4685]: I1204 06:13:50.971118 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5ec886d-9e12-44c1-97c4-7ca9597174cc","Type":"ContainerDied","Data":"ae3489794b6b378cc85c08d11244d6cbd751684b4d15244e9a0c41ddd62ca17e"} Dec 04 06:13:53 crc kubenswrapper[4685]: I1204 06:13:53.819300 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:53 crc kubenswrapper[4685]: I1204 06:13:53.828325 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:13:53 crc kubenswrapper[4685]: I1204 06:13:53.841044 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vkf95" Dec 04 06:13:55 crc kubenswrapper[4685]: I1204 06:13:55.797459 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:55 crc kubenswrapper[4685]: I1204 06:13:55.927765 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kube-api-access\") pod \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " Dec 04 06:13:55 crc kubenswrapper[4685]: I1204 06:13:55.927869 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kubelet-dir\") pod \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\" (UID: \"f5ec886d-9e12-44c1-97c4-7ca9597174cc\") " Dec 04 06:13:55 crc kubenswrapper[4685]: I1204 06:13:55.928041 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f5ec886d-9e12-44c1-97c4-7ca9597174cc" (UID: "f5ec886d-9e12-44c1-97c4-7ca9597174cc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:13:55 crc kubenswrapper[4685]: I1204 06:13:55.928194 4685 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:55 crc kubenswrapper[4685]: I1204 06:13:55.952875 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f5ec886d-9e12-44c1-97c4-7ca9597174cc" (UID: "f5ec886d-9e12-44c1-97c4-7ca9597174cc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.004622 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5ec886d-9e12-44c1-97c4-7ca9597174cc","Type":"ContainerDied","Data":"925e1035d618a292e7701412aaae02ee9a91578113aeb368277c19f8c45fb133"} Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.004681 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="925e1035d618a292e7701412aaae02ee9a91578113aeb368277c19f8c45fb133" Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.004706 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.029502 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5ec886d-9e12-44c1-97c4-7ca9597174cc-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.231593 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.236611 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06bbc9d6-e56b-4c35-bdd7-659ab3d3a870-metrics-certs\") pod \"network-metrics-daemon-55csn\" (UID: \"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870\") " pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:13:56 crc kubenswrapper[4685]: I1204 06:13:56.439246 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55csn" Dec 04 06:14:01 crc kubenswrapper[4685]: I1204 06:14:01.750579 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:14:06 crc kubenswrapper[4685]: I1204 06:14:06.121766 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:14:06 crc kubenswrapper[4685]: I1204 06:14:06.122567 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:14:09 crc kubenswrapper[4685]: E1204 06:14:09.525329 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage136495882/2\": happened during read: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 06:14:09 crc kubenswrapper[4685]: E1204 06:14:09.525981 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bthx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-g9dg5_openshift-marketplace(fc7be107-549f-4c6e-86ea-0792e58cb2ab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage136495882/2\": happened during read: context canceled" logger="UnhandledError" Dec 04 06:14:09 crc kubenswrapper[4685]: E1204 06:14:09.527153 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage136495882/2\\\": happened during read: context canceled\"" pod="openshift-marketplace/redhat-operators-g9dg5" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" Dec 04 06:14:11 crc kubenswrapper[4685]: E1204 06:14:11.125372 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 06:14:11 crc kubenswrapper[4685]: E1204 06:14:11.125648 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zz4b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-xcp78_openshift-marketplace(b2596408-3b90-481e-ad56-d5560edf0a96): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:11 crc kubenswrapper[4685]: E1204 06:14:11.127011 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-xcp78" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" Dec 04 06:14:11 crc kubenswrapper[4685]: E1204 06:14:11.927193 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-g9dg5" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.295711 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.296095 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kdn2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-872w4_openshift-marketplace(4121c660-39c8-4dd8-8a26-f40f37fb16bd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.297520 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-872w4" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.326592 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.326727 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hr6rl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2hrbf_openshift-marketplace(b1a21c41-db1c-40c9-82d1-2fb0f33f4625): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.328497 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2hrbf" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.434146 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.434286 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnjlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-47bq7_openshift-marketplace(156c8e99-d0bf-4c3f-bf15-cf0a50113b03): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:12 crc kubenswrapper[4685]: E1204 06:14:12.435614 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-47bq7" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.281055 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-47bq7" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.281365 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2hrbf" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.281430 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-xcp78" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.281469 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-872w4" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.342372 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.342528 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4vjks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-frsvx_openshift-marketplace(99e5eb96-f490-4718-8245-e7a0adee43a8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.343774 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-frsvx" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.360812 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.360946 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fpzzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6tt6d_openshift-marketplace(ca2b22c7-e7ef-445f-a910-d068f1e17575): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:13 crc kubenswrapper[4685]: E1204 06:14:13.362366 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6tt6d" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" Dec 04 06:14:14 crc kubenswrapper[4685]: I1204 06:14:14.131211 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrgc7" Dec 04 06:14:16 crc kubenswrapper[4685]: E1204 06:14:16.434160 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-frsvx" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" Dec 04 06:14:16 crc kubenswrapper[4685]: E1204 06:14:16.434203 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6tt6d" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" Dec 04 06:14:16 crc kubenswrapper[4685]: E1204 06:14:16.499952 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 06:14:16 crc kubenswrapper[4685]: E1204 06:14:16.500133 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpz72,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-892qt_openshift-marketplace(e70415f0-e1e4-4da7-a46a-ae3a01efa173): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 06:14:16 crc kubenswrapper[4685]: E1204 06:14:16.501307 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-892qt" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" Dec 04 06:14:16 crc kubenswrapper[4685]: I1204 06:14:16.507948 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 06:14:16 crc kubenswrapper[4685]: I1204 06:14:16.924334 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-55csn"] Dec 04 06:14:17 crc kubenswrapper[4685]: E1204 06:14:17.132159 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-892qt" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" Dec 04 06:14:17 crc kubenswrapper[4685]: I1204 06:14:17.135986 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-55csn" event={"ID":"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870","Type":"ContainerStarted","Data":"57dcfa895713065cdb0961332c05421083be9caee8230af10b82152d223f8227"} Dec 04 06:14:17 crc kubenswrapper[4685]: I1204 06:14:17.136049 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-55csn" event={"ID":"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870","Type":"ContainerStarted","Data":"c373ff564d855c93b66b0c2b5a8f99ab9052aec859706a61aa5b185bb4323e10"} Dec 04 06:14:18 crc kubenswrapper[4685]: I1204 06:14:18.139947 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-55csn" event={"ID":"06bbc9d6-e56b-4c35-bdd7-659ab3d3a870","Type":"ContainerStarted","Data":"20754e3fa15ad6a15192885e107a88925772a2237259ef4511a672358b54d125"} Dec 04 06:14:18 crc kubenswrapper[4685]: I1204 06:14:18.156064 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-55csn" podStartSLOduration=165.156047409 podStartE2EDuration="2m45.156047409s" podCreationTimestamp="2025-12-04 06:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:14:18.154161437 +0000 UTC m=+195.342392212" watchObservedRunningTime="2025-12-04 06:14:18.156047409 +0000 UTC m=+195.344278184" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.868287 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 06:14:20 crc kubenswrapper[4685]: E1204 06:14:20.871284 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58a129c-f8ce-4b81-8293-30e5a31373fb" containerName="pruner" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.871317 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58a129c-f8ce-4b81-8293-30e5a31373fb" containerName="pruner" Dec 04 06:14:20 crc kubenswrapper[4685]: E1204 06:14:20.871333 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ec886d-9e12-44c1-97c4-7ca9597174cc" containerName="pruner" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.871342 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ec886d-9e12-44c1-97c4-7ca9597174cc" containerName="pruner" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.872734 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ec886d-9e12-44c1-97c4-7ca9597174cc" containerName="pruner" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.872770 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b58a129c-f8ce-4b81-8293-30e5a31373fb" containerName="pruner" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.873216 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.876725 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.877483 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.878566 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.966200 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca10c436-146f-4d45-8461-382fe54049f8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:20 crc kubenswrapper[4685]: I1204 06:14:20.966271 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca10c436-146f-4d45-8461-382fe54049f8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.067657 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca10c436-146f-4d45-8461-382fe54049f8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.067754 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca10c436-146f-4d45-8461-382fe54049f8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.067795 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca10c436-146f-4d45-8461-382fe54049f8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.087114 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca10c436-146f-4d45-8461-382fe54049f8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.212315 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.616487 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 06:14:21 crc kubenswrapper[4685]: W1204 06:14:21.631959 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podca10c436_146f_4d45_8461_382fe54049f8.slice/crio-9cfbaf5b9e4a52dac66f22d59ad6d86f44c9d8fcd859c59592b1226d5ace7f2f WatchSource:0}: Error finding container 9cfbaf5b9e4a52dac66f22d59ad6d86f44c9d8fcd859c59592b1226d5ace7f2f: Status 404 returned error can't find the container with id 9cfbaf5b9e4a52dac66f22d59ad6d86f44c9d8fcd859c59592b1226d5ace7f2f Dec 04 06:14:21 crc kubenswrapper[4685]: I1204 06:14:21.836910 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rfvkg"] Dec 04 06:14:22 crc kubenswrapper[4685]: I1204 06:14:22.156609 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ca10c436-146f-4d45-8461-382fe54049f8","Type":"ContainerStarted","Data":"6ebca7407a283bee1100dc0159bc5beb1228deb13bbd57cd700f68e3f446bee5"} Dec 04 06:14:22 crc kubenswrapper[4685]: I1204 06:14:22.156896 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ca10c436-146f-4d45-8461-382fe54049f8","Type":"ContainerStarted","Data":"9cfbaf5b9e4a52dac66f22d59ad6d86f44c9d8fcd859c59592b1226d5ace7f2f"} Dec 04 06:14:22 crc kubenswrapper[4685]: I1204 06:14:22.173344 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.173323214 podStartE2EDuration="2.173323214s" podCreationTimestamp="2025-12-04 06:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:14:22.170147633 +0000 UTC m=+199.358378418" watchObservedRunningTime="2025-12-04 06:14:22.173323214 +0000 UTC m=+199.361553989" Dec 04 06:14:23 crc kubenswrapper[4685]: I1204 06:14:23.163347 4685 generic.go:334] "Generic (PLEG): container finished" podID="ca10c436-146f-4d45-8461-382fe54049f8" containerID="6ebca7407a283bee1100dc0159bc5beb1228deb13bbd57cd700f68e3f446bee5" exitCode=0 Dec 04 06:14:23 crc kubenswrapper[4685]: I1204 06:14:23.163401 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ca10c436-146f-4d45-8461-382fe54049f8","Type":"ContainerDied","Data":"6ebca7407a283bee1100dc0159bc5beb1228deb13bbd57cd700f68e3f446bee5"} Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.452810 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.512569 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca10c436-146f-4d45-8461-382fe54049f8-kube-api-access\") pod \"ca10c436-146f-4d45-8461-382fe54049f8\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.512676 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca10c436-146f-4d45-8461-382fe54049f8-kubelet-dir\") pod \"ca10c436-146f-4d45-8461-382fe54049f8\" (UID: \"ca10c436-146f-4d45-8461-382fe54049f8\") " Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.512946 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca10c436-146f-4d45-8461-382fe54049f8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ca10c436-146f-4d45-8461-382fe54049f8" (UID: "ca10c436-146f-4d45-8461-382fe54049f8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.519228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca10c436-146f-4d45-8461-382fe54049f8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ca10c436-146f-4d45-8461-382fe54049f8" (UID: "ca10c436-146f-4d45-8461-382fe54049f8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.613981 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca10c436-146f-4d45-8461-382fe54049f8-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:24 crc kubenswrapper[4685]: I1204 06:14:24.614017 4685 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca10c436-146f-4d45-8461-382fe54049f8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.175866 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ca10c436-146f-4d45-8461-382fe54049f8","Type":"ContainerDied","Data":"9cfbaf5b9e4a52dac66f22d59ad6d86f44c9d8fcd859c59592b1226d5ace7f2f"} Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.176169 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cfbaf5b9e4a52dac66f22d59ad6d86f44c9d8fcd859c59592b1226d5ace7f2f" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.175909 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.866659 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 06:14:25 crc kubenswrapper[4685]: E1204 06:14:25.867791 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca10c436-146f-4d45-8461-382fe54049f8" containerName="pruner" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.867807 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca10c436-146f-4d45-8461-382fe54049f8" containerName="pruner" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.867897 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca10c436-146f-4d45-8461-382fe54049f8" containerName="pruner" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.868234 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.871453 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.871508 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.877623 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.930133 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-var-lock\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.930212 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kube-api-access\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:25 crc kubenswrapper[4685]: I1204 06:14:25.930234 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.031857 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-var-lock\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.031940 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kube-api-access\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.031959 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.031963 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-var-lock\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.032039 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.049872 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kube-api-access\") pod \"installer-9-crc\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.183570 4685 generic.go:334] "Generic (PLEG): container finished" podID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerID="e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb" exitCode=0 Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.183675 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hrbf" event={"ID":"b1a21c41-db1c-40c9-82d1-2fb0f33f4625","Type":"ContainerDied","Data":"e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb"} Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.186100 4685 generic.go:334] "Generic (PLEG): container finished" podID="b2596408-3b90-481e-ad56-d5560edf0a96" containerID="9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae" exitCode=0 Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.186156 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcp78" event={"ID":"b2596408-3b90-481e-ad56-d5560edf0a96","Type":"ContainerDied","Data":"9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae"} Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.189524 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:14:26 crc kubenswrapper[4685]: I1204 06:14:26.575621 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 06:14:26 crc kubenswrapper[4685]: W1204 06:14:26.588665 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda2dcac27_38cb_4fd0_b9d3_54319c9d4b6e.slice/crio-0f1c2ee6b765abdf64134d2b1f9ba502fad09c6a83d61e30f53f5b1a5d90fd07 WatchSource:0}: Error finding container 0f1c2ee6b765abdf64134d2b1f9ba502fad09c6a83d61e30f53f5b1a5d90fd07: Status 404 returned error can't find the container with id 0f1c2ee6b765abdf64134d2b1f9ba502fad09c6a83d61e30f53f5b1a5d90fd07 Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.195090 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e","Type":"ContainerStarted","Data":"6f1f1d37cee5bdf0e80b2bc5633eb02f875ddb2642167c4f7fe95cdabf7c54c4"} Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.195519 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e","Type":"ContainerStarted","Data":"0f1c2ee6b765abdf64134d2b1f9ba502fad09c6a83d61e30f53f5b1a5d90fd07"} Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.199478 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcp78" event={"ID":"b2596408-3b90-481e-ad56-d5560edf0a96","Type":"ContainerStarted","Data":"52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842"} Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.201312 4685 generic.go:334] "Generic (PLEG): container finished" podID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerID="4566746d4a38c6bde1ac9a4480f1fc427eaa1741d509d53b13836e5226912637" exitCode=0 Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.201375 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872w4" event={"ID":"4121c660-39c8-4dd8-8a26-f40f37fb16bd","Type":"ContainerDied","Data":"4566746d4a38c6bde1ac9a4480f1fc427eaa1741d509d53b13836e5226912637"} Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.203895 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hrbf" event={"ID":"b1a21c41-db1c-40c9-82d1-2fb0f33f4625","Type":"ContainerStarted","Data":"aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f"} Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.214468 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.214449844 podStartE2EDuration="2.214449844s" podCreationTimestamp="2025-12-04 06:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:14:27.213112771 +0000 UTC m=+204.401343546" watchObservedRunningTime="2025-12-04 06:14:27.214449844 +0000 UTC m=+204.402680619" Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.236646 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2hrbf" podStartSLOduration=2.430155034 podStartE2EDuration="46.236628659s" podCreationTimestamp="2025-12-04 06:13:41 +0000 UTC" firstStartedPulling="2025-12-04 06:13:42.798857598 +0000 UTC m=+159.987088373" lastFinishedPulling="2025-12-04 06:14:26.605331223 +0000 UTC m=+203.793561998" observedRunningTime="2025-12-04 06:14:27.2345059 +0000 UTC m=+204.422736675" watchObservedRunningTime="2025-12-04 06:14:27.236628659 +0000 UTC m=+204.424859434" Dec 04 06:14:27 crc kubenswrapper[4685]: I1204 06:14:27.254941 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xcp78" podStartSLOduration=2.45606208 podStartE2EDuration="46.254922239s" podCreationTimestamp="2025-12-04 06:13:41 +0000 UTC" firstStartedPulling="2025-12-04 06:13:42.802443116 +0000 UTC m=+159.990673891" lastFinishedPulling="2025-12-04 06:14:26.601303275 +0000 UTC m=+203.789534050" observedRunningTime="2025-12-04 06:14:27.25367399 +0000 UTC m=+204.441904765" watchObservedRunningTime="2025-12-04 06:14:27.254922239 +0000 UTC m=+204.443153024" Dec 04 06:14:29 crc kubenswrapper[4685]: I1204 06:14:29.217472 4685 generic.go:334] "Generic (PLEG): container finished" podID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerID="47e28fdedc57f8aa959982a4a3cde03c9e091c565564492f2bf7ab608e3aa539" exitCode=0 Dec 04 06:14:29 crc kubenswrapper[4685]: I1204 06:14:29.217542 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9dg5" event={"ID":"fc7be107-549f-4c6e-86ea-0792e58cb2ab","Type":"ContainerDied","Data":"47e28fdedc57f8aa959982a4a3cde03c9e091c565564492f2bf7ab608e3aa539"} Dec 04 06:14:29 crc kubenswrapper[4685]: I1204 06:14:29.222714 4685 generic.go:334] "Generic (PLEG): container finished" podID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerID="6afdf6febf3a0ee4e5d46c37c8d53edcfb29008d694462e11bc4bcec62a52439" exitCode=0 Dec 04 06:14:29 crc kubenswrapper[4685]: I1204 06:14:29.222787 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-47bq7" event={"ID":"156c8e99-d0bf-4c3f-bf15-cf0a50113b03","Type":"ContainerDied","Data":"6afdf6febf3a0ee4e5d46c37c8d53edcfb29008d694462e11bc4bcec62a52439"} Dec 04 06:14:29 crc kubenswrapper[4685]: I1204 06:14:29.227305 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872w4" event={"ID":"4121c660-39c8-4dd8-8a26-f40f37fb16bd","Type":"ContainerStarted","Data":"d97e6433bfe318c3f82c3bacbbe8d9be37e51d1f6e5aa508decc256d08ccce67"} Dec 04 06:14:29 crc kubenswrapper[4685]: I1204 06:14:29.253740 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-872w4" podStartSLOduration=3.006746133 podStartE2EDuration="48.253714648s" podCreationTimestamp="2025-12-04 06:13:41 +0000 UTC" firstStartedPulling="2025-12-04 06:13:42.807789724 +0000 UTC m=+159.996020509" lastFinishedPulling="2025-12-04 06:14:28.054758249 +0000 UTC m=+205.242989024" observedRunningTime="2025-12-04 06:14:29.249529746 +0000 UTC m=+206.437760541" watchObservedRunningTime="2025-12-04 06:14:29.253714648 +0000 UTC m=+206.441945433" Dec 04 06:14:30 crc kubenswrapper[4685]: I1204 06:14:30.234397 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-47bq7" event={"ID":"156c8e99-d0bf-4c3f-bf15-cf0a50113b03","Type":"ContainerStarted","Data":"debf58c9ec7395cdaaad978ad213f40d00d503ddfa2dbd249f5cbfc45625910e"} Dec 04 06:14:30 crc kubenswrapper[4685]: I1204 06:14:30.236315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9dg5" event={"ID":"fc7be107-549f-4c6e-86ea-0792e58cb2ab","Type":"ContainerStarted","Data":"3e299abb3e5d4b51bdf2dfa14f705724517e2d9134dffb7f0aa0327eb7e96caa"} Dec 04 06:14:30 crc kubenswrapper[4685]: I1204 06:14:30.253362 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-47bq7" podStartSLOduration=2.440417937 podStartE2EDuration="49.253343935s" podCreationTimestamp="2025-12-04 06:13:41 +0000 UTC" firstStartedPulling="2025-12-04 06:13:42.805375193 +0000 UTC m=+159.993605958" lastFinishedPulling="2025-12-04 06:14:29.618301181 +0000 UTC m=+206.806531956" observedRunningTime="2025-12-04 06:14:30.251133815 +0000 UTC m=+207.439364590" watchObservedRunningTime="2025-12-04 06:14:30.253343935 +0000 UTC m=+207.441574710" Dec 04 06:14:30 crc kubenswrapper[4685]: I1204 06:14:30.272146 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g9dg5" podStartSLOduration=3.595135078 podStartE2EDuration="46.272125131s" podCreationTimestamp="2025-12-04 06:13:44 +0000 UTC" firstStartedPulling="2025-12-04 06:13:46.914789405 +0000 UTC m=+164.103020180" lastFinishedPulling="2025-12-04 06:14:29.591779458 +0000 UTC m=+206.780010233" observedRunningTime="2025-12-04 06:14:30.270988206 +0000 UTC m=+207.459218991" watchObservedRunningTime="2025-12-04 06:14:30.272125131 +0000 UTC m=+207.460355906" Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.243329 4685 generic.go:334] "Generic (PLEG): container finished" podID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerID="abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6" exitCode=0 Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.243418 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frsvx" event={"ID":"99e5eb96-f490-4718-8245-e7a0adee43a8","Type":"ContainerDied","Data":"abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6"} Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.246316 4685 generic.go:334] "Generic (PLEG): container finished" podID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerID="ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7" exitCode=0 Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.246357 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tt6d" event={"ID":"ca2b22c7-e7ef-445f-a910-d068f1e17575","Type":"ContainerDied","Data":"ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7"} Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.635991 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.636118 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.790090 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.847985 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.848033 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:14:31 crc kubenswrapper[4685]: I1204 06:14:31.890096 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.075717 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.075773 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.139187 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.248367 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.248499 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.256986 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerStarted","Data":"08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df"} Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.268650 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frsvx" event={"ID":"99e5eb96-f490-4718-8245-e7a0adee43a8","Type":"ContainerStarted","Data":"91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7"} Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.271542 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tt6d" event={"ID":"ca2b22c7-e7ef-445f-a910-d068f1e17575","Type":"ContainerStarted","Data":"d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721"} Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.302898 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.315639 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6tt6d" podStartSLOduration=3.359896611 podStartE2EDuration="49.315615391s" podCreationTimestamp="2025-12-04 06:13:43 +0000 UTC" firstStartedPulling="2025-12-04 06:13:45.881333047 +0000 UTC m=+163.069563822" lastFinishedPulling="2025-12-04 06:14:31.837051817 +0000 UTC m=+209.025282602" observedRunningTime="2025-12-04 06:14:32.311345465 +0000 UTC m=+209.499576240" watchObservedRunningTime="2025-12-04 06:14:32.315615391 +0000 UTC m=+209.503846166" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.316758 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.334826 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:14:32 crc kubenswrapper[4685]: I1204 06:14:32.355661 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-frsvx" podStartSLOduration=3.428921818 podStartE2EDuration="49.355641962s" podCreationTimestamp="2025-12-04 06:13:43 +0000 UTC" firstStartedPulling="2025-12-04 06:13:45.876857549 +0000 UTC m=+163.065088324" lastFinishedPulling="2025-12-04 06:14:31.803577693 +0000 UTC m=+208.991808468" observedRunningTime="2025-12-04 06:14:32.330770002 +0000 UTC m=+209.519000777" watchObservedRunningTime="2025-12-04 06:14:32.355641962 +0000 UTC m=+209.543872737" Dec 04 06:14:33 crc kubenswrapper[4685]: I1204 06:14:33.278927 4685 generic.go:334] "Generic (PLEG): container finished" podID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerID="08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df" exitCode=0 Dec 04 06:14:33 crc kubenswrapper[4685]: I1204 06:14:33.279058 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerDied","Data":"08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df"} Dec 04 06:14:33 crc kubenswrapper[4685]: I1204 06:14:33.322337 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:14:33 crc kubenswrapper[4685]: I1204 06:14:33.637118 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:14:33 crc kubenswrapper[4685]: I1204 06:14:33.637174 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:14:33 crc kubenswrapper[4685]: I1204 06:14:33.689159 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:14:34 crc kubenswrapper[4685]: I1204 06:14:34.061859 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:14:34 crc kubenswrapper[4685]: I1204 06:14:34.061910 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:14:34 crc kubenswrapper[4685]: I1204 06:14:34.112453 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:14:34 crc kubenswrapper[4685]: I1204 06:14:34.556438 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-872w4"] Dec 04 06:14:35 crc kubenswrapper[4685]: I1204 06:14:35.288502 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:14:35 crc kubenswrapper[4685]: I1204 06:14:35.288568 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:14:35 crc kubenswrapper[4685]: I1204 06:14:35.291522 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-872w4" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="registry-server" containerID="cri-o://d97e6433bfe318c3f82c3bacbbe8d9be37e51d1f6e5aa508decc256d08ccce67" gracePeriod=2 Dec 04 06:14:35 crc kubenswrapper[4685]: I1204 06:14:35.350990 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.121833 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.121898 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.121947 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.122578 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.122684 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078" gracePeriod=600 Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.298133 4685 generic.go:334] "Generic (PLEG): container finished" podID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerID="d97e6433bfe318c3f82c3bacbbe8d9be37e51d1f6e5aa508decc256d08ccce67" exitCode=0 Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.298217 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872w4" event={"ID":"4121c660-39c8-4dd8-8a26-f40f37fb16bd","Type":"ContainerDied","Data":"d97e6433bfe318c3f82c3bacbbe8d9be37e51d1f6e5aa508decc256d08ccce67"} Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.339493 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:14:36 crc kubenswrapper[4685]: I1204 06:14:36.956501 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g9dg5"] Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.306894 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078" exitCode=0 Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.307021 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078"} Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.685797 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.783287 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-catalog-content\") pod \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.783346 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdn2n\" (UniqueName: \"kubernetes.io/projected/4121c660-39c8-4dd8-8a26-f40f37fb16bd-kube-api-access-kdn2n\") pod \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.783420 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-utilities\") pod \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\" (UID: \"4121c660-39c8-4dd8-8a26-f40f37fb16bd\") " Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.784360 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-utilities" (OuterVolumeSpecName: "utilities") pod "4121c660-39c8-4dd8-8a26-f40f37fb16bd" (UID: "4121c660-39c8-4dd8-8a26-f40f37fb16bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.790030 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4121c660-39c8-4dd8-8a26-f40f37fb16bd-kube-api-access-kdn2n" (OuterVolumeSpecName: "kube-api-access-kdn2n") pod "4121c660-39c8-4dd8-8a26-f40f37fb16bd" (UID: "4121c660-39c8-4dd8-8a26-f40f37fb16bd"). InnerVolumeSpecName "kube-api-access-kdn2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.842680 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4121c660-39c8-4dd8-8a26-f40f37fb16bd" (UID: "4121c660-39c8-4dd8-8a26-f40f37fb16bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.884771 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.884822 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdn2n\" (UniqueName: \"kubernetes.io/projected/4121c660-39c8-4dd8-8a26-f40f37fb16bd-kube-api-access-kdn2n\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:37 crc kubenswrapper[4685]: I1204 06:14:37.884842 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4121c660-39c8-4dd8-8a26-f40f37fb16bd-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.318130 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872w4" event={"ID":"4121c660-39c8-4dd8-8a26-f40f37fb16bd","Type":"ContainerDied","Data":"dc38eda2f8f62624c9a39a370df183068c52a1bfecf53858170da901281d3435"} Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.318483 4685 scope.go:117] "RemoveContainer" containerID="d97e6433bfe318c3f82c3bacbbe8d9be37e51d1f6e5aa508decc256d08ccce67" Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.318240 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872w4" Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.318515 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g9dg5" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="registry-server" containerID="cri-o://3e299abb3e5d4b51bdf2dfa14f705724517e2d9134dffb7f0aa0327eb7e96caa" gracePeriod=2 Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.340185 4685 scope.go:117] "RemoveContainer" containerID="4566746d4a38c6bde1ac9a4480f1fc427eaa1741d509d53b13836e5226912637" Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.368125 4685 scope.go:117] "RemoveContainer" containerID="ef0efcc27ddd3cffc17f82771ddc7b7e1d4f2bc5e3b1711327148d13b4f3ecfc" Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.373558 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-872w4"] Dec 04 06:14:38 crc kubenswrapper[4685]: I1204 06:14:38.378292 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-872w4"] Dec 04 06:14:39 crc kubenswrapper[4685]: I1204 06:14:39.134200 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" path="/var/lib/kubelet/pods/4121c660-39c8-4dd8-8a26-f40f37fb16bd/volumes" Dec 04 06:14:39 crc kubenswrapper[4685]: I1204 06:14:39.325175 4685 generic.go:334] "Generic (PLEG): container finished" podID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerID="3e299abb3e5d4b51bdf2dfa14f705724517e2d9134dffb7f0aa0327eb7e96caa" exitCode=0 Dec 04 06:14:39 crc kubenswrapper[4685]: I1204 06:14:39.325522 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9dg5" event={"ID":"fc7be107-549f-4c6e-86ea-0792e58cb2ab","Type":"ContainerDied","Data":"3e299abb3e5d4b51bdf2dfa14f705724517e2d9134dffb7f0aa0327eb7e96caa"} Dec 04 06:14:39 crc kubenswrapper[4685]: I1204 06:14:39.327841 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerStarted","Data":"28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973"} Dec 04 06:14:39 crc kubenswrapper[4685]: I1204 06:14:39.330672 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"c538fac20a9f6d91b3fbff5866d101a5f49b5694c6b6a972d11f5cc18b5b6a38"} Dec 04 06:14:39 crc kubenswrapper[4685]: I1204 06:14:39.353527 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-892qt" podStartSLOduration=3.220111934 podStartE2EDuration="55.353500165s" podCreationTimestamp="2025-12-04 06:13:44 +0000 UTC" firstStartedPulling="2025-12-04 06:13:45.863269888 +0000 UTC m=+163.051500663" lastFinishedPulling="2025-12-04 06:14:37.996658119 +0000 UTC m=+215.184888894" observedRunningTime="2025-12-04 06:14:39.349826938 +0000 UTC m=+216.538057713" watchObservedRunningTime="2025-12-04 06:14:39.353500165 +0000 UTC m=+216.541730950" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.260263 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.317242 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-utilities\") pod \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.317300 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-catalog-content\") pod \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.317336 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bthx6\" (UniqueName: \"kubernetes.io/projected/fc7be107-549f-4c6e-86ea-0792e58cb2ab-kube-api-access-bthx6\") pod \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\" (UID: \"fc7be107-549f-4c6e-86ea-0792e58cb2ab\") " Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.318425 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-utilities" (OuterVolumeSpecName: "utilities") pod "fc7be107-549f-4c6e-86ea-0792e58cb2ab" (UID: "fc7be107-549f-4c6e-86ea-0792e58cb2ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.323799 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc7be107-549f-4c6e-86ea-0792e58cb2ab-kube-api-access-bthx6" (OuterVolumeSpecName: "kube-api-access-bthx6") pod "fc7be107-549f-4c6e-86ea-0792e58cb2ab" (UID: "fc7be107-549f-4c6e-86ea-0792e58cb2ab"). InnerVolumeSpecName "kube-api-access-bthx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.338627 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9dg5" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.339140 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9dg5" event={"ID":"fc7be107-549f-4c6e-86ea-0792e58cb2ab","Type":"ContainerDied","Data":"c0a03220f609875825cae8a56ada47d7ccbac3cce906ccaf3692f94b685b29c1"} Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.339178 4685 scope.go:117] "RemoveContainer" containerID="3e299abb3e5d4b51bdf2dfa14f705724517e2d9134dffb7f0aa0327eb7e96caa" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.359327 4685 scope.go:117] "RemoveContainer" containerID="47e28fdedc57f8aa959982a4a3cde03c9e091c565564492f2bf7ab608e3aa539" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.389441 4685 scope.go:117] "RemoveContainer" containerID="3ceae89b6a00a83bb39cae0184d3a6ce8ef1a998652db68922e3ceb0bba44b49" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.422309 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bthx6\" (UniqueName: \"kubernetes.io/projected/fc7be107-549f-4c6e-86ea-0792e58cb2ab-kube-api-access-bthx6\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.422351 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.465325 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc7be107-549f-4c6e-86ea-0792e58cb2ab" (UID: "fc7be107-549f-4c6e-86ea-0792e58cb2ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.523370 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7be107-549f-4c6e-86ea-0792e58cb2ab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.671863 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g9dg5"] Dec 04 06:14:40 crc kubenswrapper[4685]: I1204 06:14:40.675782 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g9dg5"] Dec 04 06:14:41 crc kubenswrapper[4685]: I1204 06:14:41.131868 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" path="/var/lib/kubelet/pods/fc7be107-549f-4c6e-86ea-0792e58cb2ab/volumes" Dec 04 06:14:42 crc kubenswrapper[4685]: I1204 06:14:42.116450 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:14:43 crc kubenswrapper[4685]: I1204 06:14:43.677688 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:14:43 crc kubenswrapper[4685]: I1204 06:14:43.953435 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-47bq7"] Dec 04 06:14:43 crc kubenswrapper[4685]: I1204 06:14:43.954010 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-47bq7" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="registry-server" containerID="cri-o://debf58c9ec7395cdaaad978ad213f40d00d503ddfa2dbd249f5cbfc45625910e" gracePeriod=2 Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.117902 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.366438 4685 generic.go:334] "Generic (PLEG): container finished" podID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerID="debf58c9ec7395cdaaad978ad213f40d00d503ddfa2dbd249f5cbfc45625910e" exitCode=0 Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.366484 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-47bq7" event={"ID":"156c8e99-d0bf-4c3f-bf15-cf0a50113b03","Type":"ContainerDied","Data":"debf58c9ec7395cdaaad978ad213f40d00d503ddfa2dbd249f5cbfc45625910e"} Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.510591 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.578207 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-utilities\") pod \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.578446 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-kube-api-access-vnjlp\") pod \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.578484 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-catalog-content\") pod \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\" (UID: \"156c8e99-d0bf-4c3f-bf15-cf0a50113b03\") " Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.579560 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-utilities" (OuterVolumeSpecName: "utilities") pod "156c8e99-d0bf-4c3f-bf15-cf0a50113b03" (UID: "156c8e99-d0bf-4c3f-bf15-cf0a50113b03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.584510 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-kube-api-access-vnjlp" (OuterVolumeSpecName: "kube-api-access-vnjlp") pod "156c8e99-d0bf-4c3f-bf15-cf0a50113b03" (UID: "156c8e99-d0bf-4c3f-bf15-cf0a50113b03"). InnerVolumeSpecName "kube-api-access-vnjlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.628498 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "156c8e99-d0bf-4c3f-bf15-cf0a50113b03" (UID: "156c8e99-d0bf-4c3f-bf15-cf0a50113b03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.680049 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-kube-api-access-vnjlp\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.680080 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.680093 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156c8e99-d0bf-4c3f-bf15-cf0a50113b03-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.858911 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.859272 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:14:44 crc kubenswrapper[4685]: I1204 06:14:44.904377 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.372803 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-47bq7" event={"ID":"156c8e99-d0bf-4c3f-bf15-cf0a50113b03","Type":"ContainerDied","Data":"60d82f3fd7c19e9a17a7892efd208f25bcb0b687db5d81ba70d7d162d9637d11"} Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.372859 4685 scope.go:117] "RemoveContainer" containerID="debf58c9ec7395cdaaad978ad213f40d00d503ddfa2dbd249f5cbfc45625910e" Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.372870 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-47bq7" Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.389499 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-47bq7"] Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.392084 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-47bq7"] Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.396040 4685 scope.go:117] "RemoveContainer" containerID="6afdf6febf3a0ee4e5d46c37c8d53edcfb29008d694462e11bc4bcec62a52439" Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.411468 4685 scope.go:117] "RemoveContainer" containerID="ad3d1a57e8c04ec809d7b9e4342fbf280d59c273a8169686bc647474de001ea2" Dec 04 06:14:45 crc kubenswrapper[4685]: I1204 06:14:45.415153 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:14:46 crc kubenswrapper[4685]: I1204 06:14:46.354535 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frsvx"] Dec 04 06:14:46 crc kubenswrapper[4685]: I1204 06:14:46.354754 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-frsvx" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="registry-server" containerID="cri-o://91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7" gracePeriod=2 Dec 04 06:14:46 crc kubenswrapper[4685]: I1204 06:14:46.872997 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" podUID="3ae778b1-6a99-4fec-8a14-d6d2322f18e7" containerName="oauth-openshift" containerID="cri-o://ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979" gracePeriod=15 Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.140163 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" path="/var/lib/kubelet/pods/156c8e99-d0bf-4c3f-bf15-cf0a50113b03/volumes" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.263229 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.285053 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315010 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vjks\" (UniqueName: \"kubernetes.io/projected/99e5eb96-f490-4718-8245-e7a0adee43a8-kube-api-access-4vjks\") pod \"99e5eb96-f490-4718-8245-e7a0adee43a8\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315082 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-service-ca\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315135 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-trusted-ca-bundle\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315171 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-session\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315224 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-policies\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315260 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-ocp-branding-template\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315295 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-provider-selection\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315327 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-login\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315386 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-catalog-content\") pod \"99e5eb96-f490-4718-8245-e7a0adee43a8\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315445 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc2k7\" (UniqueName: \"kubernetes.io/projected/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-kube-api-access-rc2k7\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315552 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-idp-0-file-data\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315605 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-utilities\") pod \"99e5eb96-f490-4718-8245-e7a0adee43a8\" (UID: \"99e5eb96-f490-4718-8245-e7a0adee43a8\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315647 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-serving-cert\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315692 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-router-certs\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315725 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-error\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315763 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-dir\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.315799 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-cliconfig\") pod \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\" (UID: \"3ae778b1-6a99-4fec-8a14-d6d2322f18e7\") " Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.319336 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-utilities" (OuterVolumeSpecName: "utilities") pod "99e5eb96-f490-4718-8245-e7a0adee43a8" (UID: "99e5eb96-f490-4718-8245-e7a0adee43a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.322639 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.322937 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.323149 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.323464 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.324165 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.324228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.324872 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.327769 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.331114 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.334141 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.336399 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-kube-api-access-rc2k7" (OuterVolumeSpecName: "kube-api-access-rc2k7") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "kube-api-access-rc2k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.336144 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.338442 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.338987 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3ae778b1-6a99-4fec-8a14-d6d2322f18e7" (UID: "3ae778b1-6a99-4fec-8a14-d6d2322f18e7"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.340869 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e5eb96-f490-4718-8245-e7a0adee43a8-kube-api-access-4vjks" (OuterVolumeSpecName: "kube-api-access-4vjks") pod "99e5eb96-f490-4718-8245-e7a0adee43a8" (UID: "99e5eb96-f490-4718-8245-e7a0adee43a8"). InnerVolumeSpecName "kube-api-access-4vjks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.343161 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99e5eb96-f490-4718-8245-e7a0adee43a8" (UID: "99e5eb96-f490-4718-8245-e7a0adee43a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.395296 4685 generic.go:334] "Generic (PLEG): container finished" podID="3ae778b1-6a99-4fec-8a14-d6d2322f18e7" containerID="ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979" exitCode=0 Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.395389 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.395393 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" event={"ID":"3ae778b1-6a99-4fec-8a14-d6d2322f18e7","Type":"ContainerDied","Data":"ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979"} Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.395509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rfvkg" event={"ID":"3ae778b1-6a99-4fec-8a14-d6d2322f18e7","Type":"ContainerDied","Data":"1dba8640be43b88ef2f5d02fda9d67eec072e49a48ca4b1eedc54918b95cf09c"} Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.395530 4685 scope.go:117] "RemoveContainer" containerID="ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.398254 4685 generic.go:334] "Generic (PLEG): container finished" podID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerID="91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7" exitCode=0 Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.398286 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frsvx" event={"ID":"99e5eb96-f490-4718-8245-e7a0adee43a8","Type":"ContainerDied","Data":"91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7"} Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.398327 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frsvx" event={"ID":"99e5eb96-f490-4718-8245-e7a0adee43a8","Type":"ContainerDied","Data":"c43073d9090c5dfa98f8ae01913a87e8a332a1c284eca3d4374339311f9fdfac"} Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.398438 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frsvx" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.414072 4685 scope.go:117] "RemoveContainer" containerID="ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979" Dec 04 06:14:47 crc kubenswrapper[4685]: E1204 06:14:47.414494 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979\": container with ID starting with ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979 not found: ID does not exist" containerID="ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.414550 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979"} err="failed to get container status \"ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979\": rpc error: code = NotFound desc = could not find container \"ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979\": container with ID starting with ce16a20aa8d9cb0343803d04f5961c831038cea662fc17b50103728a5cd99979 not found: ID does not exist" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.414584 4685 scope.go:117] "RemoveContainer" containerID="91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417275 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vjks\" (UniqueName: \"kubernetes.io/projected/99e5eb96-f490-4718-8245-e7a0adee43a8-kube-api-access-4vjks\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417360 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417402 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417557 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417572 4685 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417585 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417625 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417642 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417658 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417906 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc2k7\" (UniqueName: \"kubernetes.io/projected/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-kube-api-access-rc2k7\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417925 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417937 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99e5eb96-f490-4718-8245-e7a0adee43a8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417974 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.417990 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.418002 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.418016 4685 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.418028 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ae778b1-6a99-4fec-8a14-d6d2322f18e7-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.435749 4685 scope.go:117] "RemoveContainer" containerID="abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.437954 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frsvx"] Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.448741 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-frsvx"] Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.452553 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rfvkg"] Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.454918 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rfvkg"] Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.458044 4685 scope.go:117] "RemoveContainer" containerID="f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.491242 4685 scope.go:117] "RemoveContainer" containerID="91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7" Dec 04 06:14:47 crc kubenswrapper[4685]: E1204 06:14:47.491711 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7\": container with ID starting with 91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7 not found: ID does not exist" containerID="91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.491751 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7"} err="failed to get container status \"91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7\": rpc error: code = NotFound desc = could not find container \"91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7\": container with ID starting with 91277f9b5a41f1d482709f8aba805ece0d32c10b1c4f766d5a0621820f05bad7 not found: ID does not exist" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.491803 4685 scope.go:117] "RemoveContainer" containerID="abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6" Dec 04 06:14:47 crc kubenswrapper[4685]: E1204 06:14:47.492130 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6\": container with ID starting with abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6 not found: ID does not exist" containerID="abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.492161 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6"} err="failed to get container status \"abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6\": rpc error: code = NotFound desc = could not find container \"abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6\": container with ID starting with abab211339a35f2468ba0cd6fa9075f7bf58f38df5326f4005c34eea29ca1dc6 not found: ID does not exist" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.492184 4685 scope.go:117] "RemoveContainer" containerID="f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f" Dec 04 06:14:47 crc kubenswrapper[4685]: E1204 06:14:47.492546 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f\": container with ID starting with f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f not found: ID does not exist" containerID="f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f" Dec 04 06:14:47 crc kubenswrapper[4685]: I1204 06:14:47.492571 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f"} err="failed to get container status \"f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f\": rpc error: code = NotFound desc = could not find container \"f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f\": container with ID starting with f33168fb94279683810abcdf19138d3fe3ad11d741c46a1b633385a30b0aa52f not found: ID does not exist" Dec 04 06:14:49 crc kubenswrapper[4685]: I1204 06:14:49.133938 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ae778b1-6a99-4fec-8a14-d6d2322f18e7" path="/var/lib/kubelet/pods/3ae778b1-6a99-4fec-8a14-d6d2322f18e7/volumes" Dec 04 06:14:49 crc kubenswrapper[4685]: I1204 06:14:49.135185 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" path="/var/lib/kubelet/pods/99e5eb96-f490-4718-8245-e7a0adee43a8/volumes" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.561480 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v"] Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.564131 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.564391 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.564626 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ae778b1-6a99-4fec-8a14-d6d2322f18e7" containerName="oauth-openshift" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.564820 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ae778b1-6a99-4fec-8a14-d6d2322f18e7" containerName="oauth-openshift" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.564966 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.565101 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.565234 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.565365 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.565553 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.565718 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.565847 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.566177 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.566456 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.566696 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.566975 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.567218 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="extract-utilities" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.567477 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.567771 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.567921 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.568049 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.568183 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.568302 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.568462 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.568641 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: E1204 06:14:50.568809 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.568938 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="extract-content" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.569284 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4121c660-39c8-4dd8-8a26-f40f37fb16bd" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.569455 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc7be107-549f-4c6e-86ea-0792e58cb2ab" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.569614 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ae778b1-6a99-4fec-8a14-d6d2322f18e7" containerName="oauth-openshift" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.569747 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="156c8e99-d0bf-4c3f-bf15-cf0a50113b03" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.569935 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e5eb96-f490-4718-8245-e7a0adee43a8" containerName="registry-server" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.572251 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.577067 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v"] Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.580368 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.580633 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.580721 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.580858 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.581154 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.581471 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.581507 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.581603 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.582578 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.582819 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.584744 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.585119 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.590033 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.663791 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-login\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.663882 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.663940 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.663982 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-audit-dir\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664030 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-router-certs\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664096 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664133 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-audit-policies\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664181 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-service-ca\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664225 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-session\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664275 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664313 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664355 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkgs4\" (UniqueName: \"kubernetes.io/projected/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-kube-api-access-gkgs4\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664397 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-error\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.664462 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.668532 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.675399 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765192 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765237 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-audit-policies\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765260 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-service-ca\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765279 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-session\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765620 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765670 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765698 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkgs4\" (UniqueName: \"kubernetes.io/projected/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-kube-api-access-gkgs4\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765724 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-error\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765747 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765774 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-login\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765807 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765836 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765856 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-audit-dir\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.765883 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-router-certs\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.766512 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-audit-dir\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.767133 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.767151 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-audit-policies\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.767151 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-service-ca\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.768119 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.771804 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-error\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.772195 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-login\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.772594 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.773138 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.773460 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.773566 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-router-certs\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.774977 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-system-session\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.783617 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.789613 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkgs4\" (UniqueName: \"kubernetes.io/projected/4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b-kube-api-access-gkgs4\") pod \"oauth-openshift-6c5fcdcf5-txz5v\" (UID: \"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b\") " pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:50 crc kubenswrapper[4685]: I1204 06:14:50.932530 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:51 crc kubenswrapper[4685]: I1204 06:14:51.357351 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v"] Dec 04 06:14:51 crc kubenswrapper[4685]: W1204 06:14:51.366159 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c9fb9d2_77b4_47d8_9a10_b2f9c7343b9b.slice/crio-13ae2996cef62c257f08a398b0c40486c40d633e21def5cc415b9edfccdd6e4a WatchSource:0}: Error finding container 13ae2996cef62c257f08a398b0c40486c40d633e21def5cc415b9edfccdd6e4a: Status 404 returned error can't find the container with id 13ae2996cef62c257f08a398b0c40486c40d633e21def5cc415b9edfccdd6e4a Dec 04 06:14:51 crc kubenswrapper[4685]: I1204 06:14:51.427382 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" event={"ID":"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b","Type":"ContainerStarted","Data":"13ae2996cef62c257f08a398b0c40486c40d633e21def5cc415b9edfccdd6e4a"} Dec 04 06:14:52 crc kubenswrapper[4685]: I1204 06:14:52.433616 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" event={"ID":"4c9fb9d2-77b4-47d8-9a10-b2f9c7343b9b","Type":"ContainerStarted","Data":"0a1868c9d5a79fd9aeefefec4c24d531e81ad321f933d32bde4ca9018e6a1f7c"} Dec 04 06:14:52 crc kubenswrapper[4685]: I1204 06:14:52.434028 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:52 crc kubenswrapper[4685]: I1204 06:14:52.439306 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" Dec 04 06:14:52 crc kubenswrapper[4685]: I1204 06:14:52.457885 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6c5fcdcf5-txz5v" podStartSLOduration=31.45786741 podStartE2EDuration="31.45786741s" podCreationTimestamp="2025-12-04 06:14:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:14:52.456647381 +0000 UTC m=+229.644878156" watchObservedRunningTime="2025-12-04 06:14:52.45786741 +0000 UTC m=+229.646098185" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.135978 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss"] Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.138465 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.142144 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.145077 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.152033 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss"] Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.285972 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-config-volume\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.286033 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrc8j\" (UniqueName: \"kubernetes.io/projected/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-kube-api-access-vrc8j\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.286118 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-secret-volume\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.388198 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-secret-volume\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.388256 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-config-volume\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.388280 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrc8j\" (UniqueName: \"kubernetes.io/projected/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-kube-api-access-vrc8j\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.389568 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-config-volume\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.396043 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-secret-volume\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.403535 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrc8j\" (UniqueName: \"kubernetes.io/projected/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-kube-api-access-vrc8j\") pod \"collect-profiles-29413815-cktss\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.457625 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:00 crc kubenswrapper[4685]: I1204 06:15:00.650018 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss"] Dec 04 06:15:01 crc kubenswrapper[4685]: I1204 06:15:01.485396 4685 generic.go:334] "Generic (PLEG): container finished" podID="ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" containerID="73cae50a1c794021b3b7f17a6cbfd5b48162d5014ceb0f37a2b7dd40e9ab03f8" exitCode=0 Dec 04 06:15:01 crc kubenswrapper[4685]: I1204 06:15:01.485586 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" event={"ID":"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a","Type":"ContainerDied","Data":"73cae50a1c794021b3b7f17a6cbfd5b48162d5014ceb0f37a2b7dd40e9ab03f8"} Dec 04 06:15:01 crc kubenswrapper[4685]: I1204 06:15:01.485791 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" event={"ID":"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a","Type":"ContainerStarted","Data":"d3eba66c91063f6784cffc65f8f10cef6e3d536f2368a04bb51ae949a94bb0a1"} Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.809584 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.820469 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrc8j\" (UniqueName: \"kubernetes.io/projected/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-kube-api-access-vrc8j\") pod \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.820534 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-secret-volume\") pod \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.820642 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-config-volume\") pod \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\" (UID: \"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a\") " Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.822335 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-config-volume" (OuterVolumeSpecName: "config-volume") pod "ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" (UID: "ecaa27dd-6791-4c59-8c1e-7fae7c341d9a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.828252 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-kube-api-access-vrc8j" (OuterVolumeSpecName: "kube-api-access-vrc8j") pod "ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" (UID: "ecaa27dd-6791-4c59-8c1e-7fae7c341d9a"). InnerVolumeSpecName "kube-api-access-vrc8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.828620 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" (UID: "ecaa27dd-6791-4c59-8c1e-7fae7c341d9a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.921952 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.922384 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrc8j\" (UniqueName: \"kubernetes.io/projected/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-kube-api-access-vrc8j\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:02 crc kubenswrapper[4685]: I1204 06:15:02.922395 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:03 crc kubenswrapper[4685]: I1204 06:15:03.498781 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" event={"ID":"ecaa27dd-6791-4c59-8c1e-7fae7c341d9a","Type":"ContainerDied","Data":"d3eba66c91063f6784cffc65f8f10cef6e3d536f2368a04bb51ae949a94bb0a1"} Dec 04 06:15:03 crc kubenswrapper[4685]: I1204 06:15:03.498823 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3eba66c91063f6784cffc65f8f10cef6e3d536f2368a04bb51ae949a94bb0a1" Dec 04 06:15:03 crc kubenswrapper[4685]: I1204 06:15:03.498851 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.581204 4685 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.581512 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" containerName="collect-profiles" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.581528 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" containerName="collect-profiles" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.581666 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" containerName="collect-profiles" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582079 4685 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582105 4685 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582239 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582249 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582258 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582267 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582278 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582286 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582294 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582301 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582300 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582309 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582555 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582572 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582581 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 06:15:04 crc kubenswrapper[4685]: E1204 06:15:04.582595 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582604 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582597 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed" gracePeriod=15 Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582725 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582737 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582744 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582753 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582763 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.582925 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.583121 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f" gracePeriod=15 Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.583190 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6" gracePeriod=15 Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.583237 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521" gracePeriod=15 Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.583283 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e" gracePeriod=15 Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.586046 4685 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.748999 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749477 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749538 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749557 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749573 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749602 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749619 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.749659 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851029 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851082 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851129 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851150 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851167 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851177 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851215 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851218 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851233 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851254 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851697 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851703 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851725 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851774 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851796 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:04 crc kubenswrapper[4685]: I1204 06:15:04.851826 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.512078 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.513377 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.514269 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f" exitCode=0 Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.514298 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6" exitCode=0 Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.514307 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521" exitCode=0 Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.514315 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e" exitCode=2 Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.514378 4685 scope.go:117] "RemoveContainer" containerID="d257cddda201ecda2b0914bf5d1727203b2c47345c741698336036d805c97786" Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.516399 4685 generic.go:334] "Generic (PLEG): container finished" podID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" containerID="6f1f1d37cee5bdf0e80b2bc5633eb02f875ddb2642167c4f7fe95cdabf7c54c4" exitCode=0 Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.516433 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e","Type":"ContainerDied","Data":"6f1f1d37cee5bdf0e80b2bc5633eb02f875ddb2642167c4f7fe95cdabf7c54c4"} Dec 04 06:15:05 crc kubenswrapper[4685]: I1204 06:15:05.517248 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.525582 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.826987 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.828054 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979281 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kubelet-dir\") pod \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979367 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-var-lock\") pod \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979377 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" (UID: "a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979469 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kube-api-access\") pod \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\" (UID: \"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e\") " Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979477 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-var-lock" (OuterVolumeSpecName: "var-lock") pod "a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" (UID: "a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979657 4685 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.979674 4685 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:06 crc kubenswrapper[4685]: I1204 06:15:06.986634 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" (UID: "a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.080757 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.442826 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.443745 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.444479 4685 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.444855 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.534217 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e","Type":"ContainerDied","Data":"0f1c2ee6b765abdf64134d2b1f9ba502fad09c6a83d61e30f53f5b1a5d90fd07"} Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.534265 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.534276 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f1c2ee6b765abdf64134d2b1f9ba502fad09c6a83d61e30f53f5b1a5d90fd07" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.537160 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.538589 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed" exitCode=0 Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.538649 4685 scope.go:117] "RemoveContainer" containerID="78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.538899 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.539529 4685 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.539779 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.557066 4685 scope.go:117] "RemoveContainer" containerID="e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.576232 4685 scope.go:117] "RemoveContainer" containerID="1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.585736 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.585921 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.586045 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.586380 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.586441 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.586471 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.593223 4685 scope.go:117] "RemoveContainer" containerID="2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.607372 4685 scope.go:117] "RemoveContainer" containerID="3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.620494 4685 scope.go:117] "RemoveContainer" containerID="220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.641131 4685 scope.go:117] "RemoveContainer" containerID="78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f" Dec 04 06:15:07 crc kubenswrapper[4685]: E1204 06:15:07.643687 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\": container with ID starting with 78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f not found: ID does not exist" containerID="78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.643717 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f"} err="failed to get container status \"78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\": rpc error: code = NotFound desc = could not find container \"78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f\": container with ID starting with 78e22ce6f896849ac403a3bbb5e5e34d2b8a5b8fd0731a4a94dd448fe9004f9f not found: ID does not exist" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.643741 4685 scope.go:117] "RemoveContainer" containerID="e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6" Dec 04 06:15:07 crc kubenswrapper[4685]: E1204 06:15:07.644094 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\": container with ID starting with e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6 not found: ID does not exist" containerID="e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.644118 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6"} err="failed to get container status \"e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\": rpc error: code = NotFound desc = could not find container \"e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6\": container with ID starting with e83b5b0b89d0926f413de304be4ad1209ac4781761af9ddc1662b2b9b172fef6 not found: ID does not exist" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.644139 4685 scope.go:117] "RemoveContainer" containerID="1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521" Dec 04 06:15:07 crc kubenswrapper[4685]: E1204 06:15:07.644577 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\": container with ID starting with 1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521 not found: ID does not exist" containerID="1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.644599 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521"} err="failed to get container status \"1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\": rpc error: code = NotFound desc = could not find container \"1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521\": container with ID starting with 1508c097f3d361c9b6b734902eb3812e8be3e414b1a862f7e09ee45f828f9521 not found: ID does not exist" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.644612 4685 scope.go:117] "RemoveContainer" containerID="2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e" Dec 04 06:15:07 crc kubenswrapper[4685]: E1204 06:15:07.644926 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\": container with ID starting with 2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e not found: ID does not exist" containerID="2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.644946 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e"} err="failed to get container status \"2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\": rpc error: code = NotFound desc = could not find container \"2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e\": container with ID starting with 2033c158958e9002555582fe64fbee543d618d52b3ec98a04a47dda34fc9d20e not found: ID does not exist" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.644958 4685 scope.go:117] "RemoveContainer" containerID="3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed" Dec 04 06:15:07 crc kubenswrapper[4685]: E1204 06:15:07.645232 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\": container with ID starting with 3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed not found: ID does not exist" containerID="3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.645268 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed"} err="failed to get container status \"3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\": rpc error: code = NotFound desc = could not find container \"3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed\": container with ID starting with 3e608951141e8e0456458ac0b8a79ba9ca6b9b676a304c51a81c69503755a4ed not found: ID does not exist" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.645376 4685 scope.go:117] "RemoveContainer" containerID="220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624" Dec 04 06:15:07 crc kubenswrapper[4685]: E1204 06:15:07.645747 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\": container with ID starting with 220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624 not found: ID does not exist" containerID="220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.645801 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624"} err="failed to get container status \"220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\": rpc error: code = NotFound desc = could not find container \"220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624\": container with ID starting with 220b7000d271631567e50985a50ddb0748ee02a9c79f572e99df5fb455bb2624 not found: ID does not exist" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.688225 4685 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.688462 4685 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.688475 4685 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.857374 4685 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:07 crc kubenswrapper[4685]: I1204 06:15:07.857750 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:09 crc kubenswrapper[4685]: I1204 06:15:09.132308 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 06:15:09 crc kubenswrapper[4685]: E1204 06:15:09.612971 4685 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:09 crc kubenswrapper[4685]: I1204 06:15:09.613481 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:09 crc kubenswrapper[4685]: E1204 06:15:09.635205 4685 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dee8af811ed62 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 06:15:09.63392445 +0000 UTC m=+246.822155225,LastTimestamp:2025-12-04 06:15:09.63392445 +0000 UTC m=+246.822155225,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 06:15:10 crc kubenswrapper[4685]: I1204 06:15:10.556563 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42"} Dec 04 06:15:10 crc kubenswrapper[4685]: I1204 06:15:10.556875 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2b7b182d689de43ec27e78089387271c477898d77b3c517d8087785b33ab3061"} Dec 04 06:15:10 crc kubenswrapper[4685]: I1204 06:15:10.557889 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.558122 4685 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.959648 4685 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.960257 4685 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.960721 4685 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.961014 4685 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.961387 4685 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:10 crc kubenswrapper[4685]: I1204 06:15:10.961458 4685 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 06:15:10 crc kubenswrapper[4685]: E1204 06:15:10.961790 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="200ms" Dec 04 06:15:11 crc kubenswrapper[4685]: E1204 06:15:11.162786 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="400ms" Dec 04 06:15:11 crc kubenswrapper[4685]: E1204 06:15:11.563722 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="800ms" Dec 04 06:15:12 crc kubenswrapper[4685]: E1204 06:15:12.365045 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="1.6s" Dec 04 06:15:13 crc kubenswrapper[4685]: I1204 06:15:13.128598 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:13 crc kubenswrapper[4685]: E1204 06:15:13.128643 4685 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dee8af811ed62 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 06:15:09.63392445 +0000 UTC m=+246.822155225,LastTimestamp:2025-12-04 06:15:09.63392445 +0000 UTC m=+246.822155225,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 06:15:13 crc kubenswrapper[4685]: E1204 06:15:13.966946 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="3.2s" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.125320 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.126109 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.141524 4685 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.141870 4685 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:15 crc kubenswrapper[4685]: E1204 06:15:15.142276 4685 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.143007 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.590650 4685 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="22198bbc54b26c004f67d7c55ca4a34cd12586cf36e415e9409ddfed69e1b8c2" exitCode=0 Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.590770 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"22198bbc54b26c004f67d7c55ca4a34cd12586cf36e415e9409ddfed69e1b8c2"} Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.591085 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"be6ac949c6deb5a2c2146968ca67c4f2ccb6bddb250abece9aca59baae6ed41d"} Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.591353 4685 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.591366 4685 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:15 crc kubenswrapper[4685]: I1204 06:15:15.592582 4685 status_manager.go:851] "Failed to get status for pod" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Dec 04 06:15:15 crc kubenswrapper[4685]: E1204 06:15:15.592579 4685 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:16 crc kubenswrapper[4685]: I1204 06:15:16.626229 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"79d0f5f4f8f159bd95773d04bd9bd1c299e92b9617c3953d41215c85dd107c9b"} Dec 04 06:15:16 crc kubenswrapper[4685]: I1204 06:15:16.626994 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b80f172352169b6baed15d525ac021af40c4e3a1cd2dc471e511c7c8bd9804eb"} Dec 04 06:15:16 crc kubenswrapper[4685]: I1204 06:15:16.627015 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"821126ede2e4918aa2e62e228590dc5c358350fb583e26a689ca9e4f99174be1"} Dec 04 06:15:16 crc kubenswrapper[4685]: I1204 06:15:16.627047 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"728e20664e9ad972b7f8a1270da482f586e44c29de31c7cb64941fbbbcd016ba"} Dec 04 06:15:17 crc kubenswrapper[4685]: I1204 06:15:17.634267 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"67e6d4160976a556bb3c1ff69353bd74484f06ba29d4064683461fa48caea4bb"} Dec 04 06:15:17 crc kubenswrapper[4685]: I1204 06:15:17.634459 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:17 crc kubenswrapper[4685]: I1204 06:15:17.634639 4685 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:17 crc kubenswrapper[4685]: I1204 06:15:17.634672 4685 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:18 crc kubenswrapper[4685]: I1204 06:15:18.642522 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 06:15:18 crc kubenswrapper[4685]: I1204 06:15:18.642837 4685 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3" exitCode=1 Dec 04 06:15:18 crc kubenswrapper[4685]: I1204 06:15:18.642864 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3"} Dec 04 06:15:18 crc kubenswrapper[4685]: I1204 06:15:18.643291 4685 scope.go:117] "RemoveContainer" containerID="801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3" Dec 04 06:15:19 crc kubenswrapper[4685]: I1204 06:15:19.653444 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 06:15:19 crc kubenswrapper[4685]: I1204 06:15:19.653508 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1d5cac1482b7d8a618a67d1affcc35dda0d5481c7a627db3d606d899a6bd9bb0"} Dec 04 06:15:20 crc kubenswrapper[4685]: I1204 06:15:20.143790 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:20 crc kubenswrapper[4685]: I1204 06:15:20.144193 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:20 crc kubenswrapper[4685]: I1204 06:15:20.149035 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:22 crc kubenswrapper[4685]: I1204 06:15:22.643581 4685 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:22 crc kubenswrapper[4685]: I1204 06:15:22.669684 4685 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:22 crc kubenswrapper[4685]: I1204 06:15:22.669727 4685 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:22 crc kubenswrapper[4685]: I1204 06:15:22.676552 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:23 crc kubenswrapper[4685]: I1204 06:15:23.133961 4685 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3270db80-baf8-4ecd-8499-206bb146adca" Dec 04 06:15:23 crc kubenswrapper[4685]: I1204 06:15:23.573831 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:15:23 crc kubenswrapper[4685]: I1204 06:15:23.675503 4685 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:23 crc kubenswrapper[4685]: I1204 06:15:23.678083 4685 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="546be9fc-d625-4c4b-a8e1-9fc0ff588b2c" Dec 04 06:15:23 crc kubenswrapper[4685]: I1204 06:15:23.680108 4685 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3270db80-baf8-4ecd-8499-206bb146adca" Dec 04 06:15:24 crc kubenswrapper[4685]: I1204 06:15:24.112044 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:15:24 crc kubenswrapper[4685]: I1204 06:15:24.112298 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 06:15:24 crc kubenswrapper[4685]: I1204 06:15:24.112341 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.057392 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.115935 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.134929 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.136206 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.263377 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.325580 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.341785 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.428207 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.449710 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.476475 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.479882 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 06:15:29 crc kubenswrapper[4685]: I1204 06:15:29.648904 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.149368 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.157449 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.345728 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.591123 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.658682 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.858449 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.934983 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 06:15:30 crc kubenswrapper[4685]: I1204 06:15:30.940621 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 06:15:31 crc kubenswrapper[4685]: I1204 06:15:31.306369 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 06:15:31 crc kubenswrapper[4685]: I1204 06:15:31.501653 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 06:15:31 crc kubenswrapper[4685]: I1204 06:15:31.674710 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 06:15:31 crc kubenswrapper[4685]: I1204 06:15:31.713398 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 06:15:31 crc kubenswrapper[4685]: I1204 06:15:31.899606 4685 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 06:15:31 crc kubenswrapper[4685]: I1204 06:15:31.937393 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.133660 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.271565 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.317802 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.403573 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.466490 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.544210 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.722327 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.746109 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 06:15:32 crc kubenswrapper[4685]: I1204 06:15:32.863894 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.081548 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.095106 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.135053 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.498149 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.505970 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.715716 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.752150 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 06:15:33 crc kubenswrapper[4685]: I1204 06:15:33.788529 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.008561 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.062018 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.075956 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.112711 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.112790 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.549533 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.621299 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.736564 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 06:15:34 crc kubenswrapper[4685]: I1204 06:15:34.848177 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 06:15:35 crc kubenswrapper[4685]: I1204 06:15:35.038585 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 06:15:35 crc kubenswrapper[4685]: I1204 06:15:35.458911 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 06:15:35 crc kubenswrapper[4685]: I1204 06:15:35.513578 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 06:15:35 crc kubenswrapper[4685]: I1204 06:15:35.927199 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 06:15:36 crc kubenswrapper[4685]: I1204 06:15:36.095348 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 06:15:36 crc kubenswrapper[4685]: I1204 06:15:36.348357 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 06:15:36 crc kubenswrapper[4685]: I1204 06:15:36.393929 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 06:15:36 crc kubenswrapper[4685]: I1204 06:15:36.495710 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.194478 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.236006 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.240864 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.302488 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.349636 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.375436 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.411218 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.493930 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.552833 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.554427 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.680548 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.724196 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 06:15:37 crc kubenswrapper[4685]: I1204 06:15:37.904197 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.051528 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.167496 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.186921 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.195052 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.327249 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.329210 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.394164 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.545836 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.772888 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.807442 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.956723 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 06:15:38 crc kubenswrapper[4685]: I1204 06:15:38.988068 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.336852 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.510586 4685 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.547445 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.584491 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.641070 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.708366 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.708523 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.828596 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.829164 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 06:15:39 crc kubenswrapper[4685]: I1204 06:15:39.864927 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.092758 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.093078 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.340590 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.424247 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.515678 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.573960 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.740559 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.811915 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.945777 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 06:15:40 crc kubenswrapper[4685]: I1204 06:15:40.953393 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.004214 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.137646 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.295507 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.328238 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.333936 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.358717 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.358848 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.378706 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.379711 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.390556 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.393063 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.414352 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.442633 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.493454 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.513438 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.530012 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.530086 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.548827 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.571134 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.652423 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.813587 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.826934 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.885804 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.895287 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.915378 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 06:15:41 crc kubenswrapper[4685]: I1204 06:15:41.970246 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.082204 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.163922 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.204804 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.296724 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.314633 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.337930 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.417267 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.424762 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.429262 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.512105 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.539039 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.584633 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.604077 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.694283 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.765177 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.777185 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.877301 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.927720 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 06:15:42 crc kubenswrapper[4685]: I1204 06:15:42.942583 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.008974 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.039397 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.053730 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.062611 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.088863 4685 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.094652 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.094716 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.106878 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.106895 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.125650 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.125633693 podStartE2EDuration="21.125633693s" podCreationTimestamp="2025-12-04 06:15:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:15:43.123331618 +0000 UTC m=+280.311562393" watchObservedRunningTime="2025-12-04 06:15:43.125633693 +0000 UTC m=+280.313864468" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.163256 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xcp78"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.163618 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xcp78" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="registry-server" containerID="cri-o://52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842" gracePeriod=30 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.168510 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2hrbf"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.169516 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2hrbf" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="registry-server" containerID="cri-o://aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f" gracePeriod=30 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.173515 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crn6w"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.174052 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" containerName="marketplace-operator" containerID="cri-o://7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824" gracePeriod=30 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.179097 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tt6d"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.179223 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.179265 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.179544 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6tt6d" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="registry-server" containerID="cri-o://d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" gracePeriod=30 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.186704 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-892qt"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.191864 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-892qt" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="registry-server" containerID="cri-o://28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973" gracePeriod=30 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.254899 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.302191 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.312285 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.372605 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.385274 4685 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.458622 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.587024 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.599605 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.600112 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.637184 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721 is running failed: container process not found" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.637584 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.638431 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721 is running failed: container process not found" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.639535 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721 is running failed: container process not found" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.639616 4685 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-6tt6d" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="registry-server" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.682999 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-utilities\") pod \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.683075 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr6rl\" (UniqueName: \"kubernetes.io/projected/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-kube-api-access-hr6rl\") pod \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.683236 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-catalog-content\") pod \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\" (UID: \"b1a21c41-db1c-40c9-82d1-2fb0f33f4625\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.684340 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-utilities" (OuterVolumeSpecName: "utilities") pod "b1a21c41-db1c-40c9-82d1-2fb0f33f4625" (UID: "b1a21c41-db1c-40c9-82d1-2fb0f33f4625"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.690682 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-kube-api-access-hr6rl" (OuterVolumeSpecName: "kube-api-access-hr6rl") pod "b1a21c41-db1c-40c9-82d1-2fb0f33f4625" (UID: "b1a21c41-db1c-40c9-82d1-2fb0f33f4625"). InnerVolumeSpecName "kube-api-access-hr6rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.710217 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.726206 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.735040 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.738068 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.738614 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.746988 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1a21c41-db1c-40c9-82d1-2fb0f33f4625" (UID: "b1a21c41-db1c-40c9-82d1-2fb0f33f4625"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.750692 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784299 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-utilities\") pod \"b2596408-3b90-481e-ad56-d5560edf0a96\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784370 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz4b6\" (UniqueName: \"kubernetes.io/projected/b2596408-3b90-481e-ad56-d5560edf0a96-kube-api-access-zz4b6\") pod \"b2596408-3b90-481e-ad56-d5560edf0a96\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784447 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-utilities\") pod \"ca2b22c7-e7ef-445f-a910-d068f1e17575\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784475 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvbsx\" (UniqueName: \"kubernetes.io/projected/7736df92-b993-435a-ad74-4c2edd644ca3-kube-api-access-nvbsx\") pod \"7736df92-b993-435a-ad74-4c2edd644ca3\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784536 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-trusted-ca\") pod \"7736df92-b993-435a-ad74-4c2edd644ca3\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784559 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-operator-metrics\") pod \"7736df92-b993-435a-ad74-4c2edd644ca3\" (UID: \"7736df92-b993-435a-ad74-4c2edd644ca3\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784623 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpzzx\" (UniqueName: \"kubernetes.io/projected/ca2b22c7-e7ef-445f-a910-d068f1e17575-kube-api-access-fpzzx\") pod \"ca2b22c7-e7ef-445f-a910-d068f1e17575\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784646 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-catalog-content\") pod \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784675 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpz72\" (UniqueName: \"kubernetes.io/projected/e70415f0-e1e4-4da7-a46a-ae3a01efa173-kube-api-access-xpz72\") pod \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784710 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-catalog-content\") pod \"b2596408-3b90-481e-ad56-d5560edf0a96\" (UID: \"b2596408-3b90-481e-ad56-d5560edf0a96\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784732 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-utilities\") pod \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\" (UID: \"e70415f0-e1e4-4da7-a46a-ae3a01efa173\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.784768 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-catalog-content\") pod \"ca2b22c7-e7ef-445f-a910-d068f1e17575\" (UID: \"ca2b22c7-e7ef-445f-a910-d068f1e17575\") " Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.785082 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.785098 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.785113 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr6rl\" (UniqueName: \"kubernetes.io/projected/b1a21c41-db1c-40c9-82d1-2fb0f33f4625-kube-api-access-hr6rl\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.786289 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-utilities" (OuterVolumeSpecName: "utilities") pod "b2596408-3b90-481e-ad56-d5560edf0a96" (UID: "b2596408-3b90-481e-ad56-d5560edf0a96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.786850 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-utilities" (OuterVolumeSpecName: "utilities") pod "e70415f0-e1e4-4da7-a46a-ae3a01efa173" (UID: "e70415f0-e1e4-4da7-a46a-ae3a01efa173"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.787156 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7736df92-b993-435a-ad74-4c2edd644ca3" (UID: "7736df92-b993-435a-ad74-4c2edd644ca3"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.788736 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7736df92-b993-435a-ad74-4c2edd644ca3" (UID: "7736df92-b993-435a-ad74-4c2edd644ca3"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.789184 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2596408-3b90-481e-ad56-d5560edf0a96-kube-api-access-zz4b6" (OuterVolumeSpecName: "kube-api-access-zz4b6") pod "b2596408-3b90-481e-ad56-d5560edf0a96" (UID: "b2596408-3b90-481e-ad56-d5560edf0a96"). InnerVolumeSpecName "kube-api-access-zz4b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.790613 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2b22c7-e7ef-445f-a910-d068f1e17575-kube-api-access-fpzzx" (OuterVolumeSpecName: "kube-api-access-fpzzx") pod "ca2b22c7-e7ef-445f-a910-d068f1e17575" (UID: "ca2b22c7-e7ef-445f-a910-d068f1e17575"). InnerVolumeSpecName "kube-api-access-fpzzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.790685 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70415f0-e1e4-4da7-a46a-ae3a01efa173-kube-api-access-xpz72" (OuterVolumeSpecName: "kube-api-access-xpz72") pod "e70415f0-e1e4-4da7-a46a-ae3a01efa173" (UID: "e70415f0-e1e4-4da7-a46a-ae3a01efa173"). InnerVolumeSpecName "kube-api-access-xpz72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.790785 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7736df92-b993-435a-ad74-4c2edd644ca3-kube-api-access-nvbsx" (OuterVolumeSpecName: "kube-api-access-nvbsx") pod "7736df92-b993-435a-ad74-4c2edd644ca3" (UID: "7736df92-b993-435a-ad74-4c2edd644ca3"). InnerVolumeSpecName "kube-api-access-nvbsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.792695 4685 generic.go:334] "Generic (PLEG): container finished" podID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" exitCode=0 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.792765 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tt6d" event={"ID":"ca2b22c7-e7ef-445f-a910-d068f1e17575","Type":"ContainerDied","Data":"d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.792799 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tt6d" event={"ID":"ca2b22c7-e7ef-445f-a910-d068f1e17575","Type":"ContainerDied","Data":"c31577c6525d1724d066c6b34dec76f775bc66911f4e26af918e3d8672598578"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.792820 4685 scope.go:117] "RemoveContainer" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.792960 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tt6d" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.797625 4685 generic.go:334] "Generic (PLEG): container finished" podID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerID="aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f" exitCode=0 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.797682 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hrbf" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.797691 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hrbf" event={"ID":"b1a21c41-db1c-40c9-82d1-2fb0f33f4625","Type":"ContainerDied","Data":"aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.797712 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hrbf" event={"ID":"b1a21c41-db1c-40c9-82d1-2fb0f33f4625","Type":"ContainerDied","Data":"aec6c76df4f77c213ea64af7070859b4ccfeddf45a8a65361e319c6da6f63d65"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.800119 4685 generic.go:334] "Generic (PLEG): container finished" podID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerID="28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973" exitCode=0 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.800182 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerDied","Data":"28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.800206 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-892qt" event={"ID":"e70415f0-e1e4-4da7-a46a-ae3a01efa173","Type":"ContainerDied","Data":"69196d35d40f446733c9567195b0bc8a4f157a42c240811999b54ed27635d3fc"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.800277 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-892qt" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.800514 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-utilities" (OuterVolumeSpecName: "utilities") pod "ca2b22c7-e7ef-445f-a910-d068f1e17575" (UID: "ca2b22c7-e7ef-445f-a910-d068f1e17575"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.805418 4685 generic.go:334] "Generic (PLEG): container finished" podID="b2596408-3b90-481e-ad56-d5560edf0a96" containerID="52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842" exitCode=0 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.805537 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcp78" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.805537 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcp78" event={"ID":"b2596408-3b90-481e-ad56-d5560edf0a96","Type":"ContainerDied","Data":"52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.805637 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcp78" event={"ID":"b2596408-3b90-481e-ad56-d5560edf0a96","Type":"ContainerDied","Data":"d3314796cffe16cc933e05695536cae8b78c7c50ba2b4ece75c0fa8f92fb483e"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.809297 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.809965 4685 generic.go:334] "Generic (PLEG): container finished" podID="7736df92-b993-435a-ad74-4c2edd644ca3" containerID="7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824" exitCode=0 Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.810385 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.810484 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" event={"ID":"7736df92-b993-435a-ad74-4c2edd644ca3","Type":"ContainerDied","Data":"7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.810535 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-crn6w" event={"ID":"7736df92-b993-435a-ad74-4c2edd644ca3","Type":"ContainerDied","Data":"587658f83a4372449257a5d4318019a90d96e2154c4dcaa48742344d95d3fd87"} Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.815065 4685 scope.go:117] "RemoveContainer" containerID="ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.831498 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2hrbf"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.835435 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2hrbf"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.839588 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca2b22c7-e7ef-445f-a910-d068f1e17575" (UID: "ca2b22c7-e7ef-445f-a910-d068f1e17575"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.840670 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crn6w"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.843099 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crn6w"] Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.853104 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2596408-3b90-481e-ad56-d5560edf0a96" (UID: "b2596408-3b90-481e-ad56-d5560edf0a96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.859801 4685 scope.go:117] "RemoveContainer" containerID="2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.874868 4685 scope.go:117] "RemoveContainer" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.875440 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721\": container with ID starting with d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721 not found: ID does not exist" containerID="d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.875488 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721"} err="failed to get container status \"d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721\": rpc error: code = NotFound desc = could not find container \"d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721\": container with ID starting with d23cbaaeab7fcbb35cc794e5079e26c99c3a00ed2427d5ab17abdd566fd49721 not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.875516 4685 scope.go:117] "RemoveContainer" containerID="ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.876066 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7\": container with ID starting with ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7 not found: ID does not exist" containerID="ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.876123 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7"} err="failed to get container status \"ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7\": rpc error: code = NotFound desc = could not find container \"ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7\": container with ID starting with ff8bee48cad950893c8c37ed706b04a5c4b7aef66607e9d46e3fabc2bd9556e7 not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.876158 4685 scope.go:117] "RemoveContainer" containerID="2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.876552 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688\": container with ID starting with 2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688 not found: ID does not exist" containerID="2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.876609 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688"} err="failed to get container status \"2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688\": rpc error: code = NotFound desc = could not find container \"2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688\": container with ID starting with 2ff9368f3cafa90ed639ca78a0d9fa4e821bd41765830d1341c97546c1259688 not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.876630 4685 scope.go:117] "RemoveContainer" containerID="aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886103 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz4b6\" (UniqueName: \"kubernetes.io/projected/b2596408-3b90-481e-ad56-d5560edf0a96-kube-api-access-zz4b6\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886437 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886540 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvbsx\" (UniqueName: \"kubernetes.io/projected/7736df92-b993-435a-ad74-4c2edd644ca3-kube-api-access-nvbsx\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886614 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886696 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7736df92-b993-435a-ad74-4c2edd644ca3-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886768 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpzzx\" (UniqueName: \"kubernetes.io/projected/ca2b22c7-e7ef-445f-a910-d068f1e17575-kube-api-access-fpzzx\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886862 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpz72\" (UniqueName: \"kubernetes.io/projected/e70415f0-e1e4-4da7-a46a-ae3a01efa173-kube-api-access-xpz72\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.886970 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.887043 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.887109 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2b22c7-e7ef-445f-a910-d068f1e17575-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.887169 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2596408-3b90-481e-ad56-d5560edf0a96-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.891020 4685 scope.go:117] "RemoveContainer" containerID="e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.908923 4685 scope.go:117] "RemoveContainer" containerID="b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.921151 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e70415f0-e1e4-4da7-a46a-ae3a01efa173" (UID: "e70415f0-e1e4-4da7-a46a-ae3a01efa173"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.923557 4685 scope.go:117] "RemoveContainer" containerID="aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.924047 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f\": container with ID starting with aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f not found: ID does not exist" containerID="aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.924079 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f"} err="failed to get container status \"aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f\": rpc error: code = NotFound desc = could not find container \"aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f\": container with ID starting with aff18ed5cb162f6666982c3b6c8bc2642325b987257f188d7efc56b978c73b0f not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.924099 4685 scope.go:117] "RemoveContainer" containerID="e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.924514 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb\": container with ID starting with e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb not found: ID does not exist" containerID="e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.924536 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb"} err="failed to get container status \"e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb\": rpc error: code = NotFound desc = could not find container \"e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb\": container with ID starting with e7a2bc3b7b99f3fd232fd843f8b3b38d5fbaa7ba3123741ef7558117c7cc35bb not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.924549 4685 scope.go:117] "RemoveContainer" containerID="b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.924819 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d\": container with ID starting with b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d not found: ID does not exist" containerID="b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.924839 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d"} err="failed to get container status \"b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d\": rpc error: code = NotFound desc = could not find container \"b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d\": container with ID starting with b0e0746fa7ad5e5f9b3403d8ba9343d887ace5990f32bbcddb3e79a9a986947d not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.924852 4685 scope.go:117] "RemoveContainer" containerID="28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.940202 4685 scope.go:117] "RemoveContainer" containerID="08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.958723 4685 scope.go:117] "RemoveContainer" containerID="b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.976557 4685 scope.go:117] "RemoveContainer" containerID="28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.977546 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973\": container with ID starting with 28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973 not found: ID does not exist" containerID="28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.977589 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973"} err="failed to get container status \"28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973\": rpc error: code = NotFound desc = could not find container \"28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973\": container with ID starting with 28cf1cf80e09ffd6ff6b5e98e615af84c94f744f8109cd0148283e73022a6973 not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.977620 4685 scope.go:117] "RemoveContainer" containerID="08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.977962 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df\": container with ID starting with 08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df not found: ID does not exist" containerID="08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.977983 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df"} err="failed to get container status \"08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df\": rpc error: code = NotFound desc = could not find container \"08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df\": container with ID starting with 08c7f1202bf008360ebb2d0b821f34b1cc957cc798527ca44d5b8d9171baa3df not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.978199 4685 scope.go:117] "RemoveContainer" containerID="b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a" Dec 04 06:15:43 crc kubenswrapper[4685]: E1204 06:15:43.978555 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a\": container with ID starting with b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a not found: ID does not exist" containerID="b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.978586 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a"} err="failed to get container status \"b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a\": rpc error: code = NotFound desc = could not find container \"b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a\": container with ID starting with b5e28f43de1f26e44d3b390df025713b3c08a42a1c06ab5c87e24fbfc5872d3a not found: ID does not exist" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.978603 4685 scope.go:117] "RemoveContainer" containerID="52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.987867 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70415f0-e1e4-4da7-a46a-ae3a01efa173-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:43 crc kubenswrapper[4685]: I1204 06:15:43.993178 4685 scope.go:117] "RemoveContainer" containerID="9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.002339 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.002509 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.010456 4685 scope.go:117] "RemoveContainer" containerID="f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.027812 4685 scope.go:117] "RemoveContainer" containerID="52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842" Dec 04 06:15:44 crc kubenswrapper[4685]: E1204 06:15:44.028324 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842\": container with ID starting with 52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842 not found: ID does not exist" containerID="52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.028354 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842"} err="failed to get container status \"52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842\": rpc error: code = NotFound desc = could not find container \"52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842\": container with ID starting with 52e65c02bb51048fdf2ee91405110dff5055666c11a9a63cd47840090438d842 not found: ID does not exist" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.028374 4685 scope.go:117] "RemoveContainer" containerID="9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae" Dec 04 06:15:44 crc kubenswrapper[4685]: E1204 06:15:44.028641 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae\": container with ID starting with 9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae not found: ID does not exist" containerID="9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.028673 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae"} err="failed to get container status \"9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae\": rpc error: code = NotFound desc = could not find container \"9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae\": container with ID starting with 9454a7b4cbc6c55811ccd2fda73fa917d2d20837e2928c30532573f1dfedd0ae not found: ID does not exist" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.028691 4685 scope.go:117] "RemoveContainer" containerID="f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196" Dec 04 06:15:44 crc kubenswrapper[4685]: E1204 06:15:44.028912 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196\": container with ID starting with f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196 not found: ID does not exist" containerID="f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.028934 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196"} err="failed to get container status \"f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196\": rpc error: code = NotFound desc = could not find container \"f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196\": container with ID starting with f91b27367cc3babd8a9ec64b99d7c92d7fb6cf2d3a27bfcda768e596d1b2a196 not found: ID does not exist" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.028946 4685 scope.go:117] "RemoveContainer" containerID="7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.042059 4685 scope.go:117] "RemoveContainer" containerID="7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824" Dec 04 06:15:44 crc kubenswrapper[4685]: E1204 06:15:44.042456 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824\": container with ID starting with 7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824 not found: ID does not exist" containerID="7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.042498 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824"} err="failed to get container status \"7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824\": rpc error: code = NotFound desc = could not find container \"7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824\": container with ID starting with 7fb081b91ae9717d5df309ffd5ca14b17c7a842b583e9780d0f77bd811f83824 not found: ID does not exist" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.075157 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.112820 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.112895 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.113162 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.120255 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"1d5cac1482b7d8a618a67d1affcc35dda0d5481c7a627db3d606d899a6bd9bb0"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.120397 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://1d5cac1482b7d8a618a67d1affcc35dda0d5481c7a627db3d606d899a6bd9bb0" gracePeriod=30 Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.128722 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tt6d"] Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.134769 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tt6d"] Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.149603 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-892qt"] Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.154253 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-892qt"] Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.167547 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xcp78"] Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.171144 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xcp78"] Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.183158 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.356899 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.449730 4685 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.472350 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.538054 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.540464 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.710196 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.839010 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.847236 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.916704 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 06:15:44 crc kubenswrapper[4685]: I1204 06:15:44.943654 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.079877 4685 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.080140 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42" gracePeriod=5 Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.095705 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.134088 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" path="/var/lib/kubelet/pods/7736df92-b993-435a-ad74-4c2edd644ca3/volumes" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.134741 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" path="/var/lib/kubelet/pods/b1a21c41-db1c-40c9-82d1-2fb0f33f4625/volumes" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.135517 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" path="/var/lib/kubelet/pods/b2596408-3b90-481e-ad56-d5560edf0a96/volumes" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.136882 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" path="/var/lib/kubelet/pods/ca2b22c7-e7ef-445f-a910-d068f1e17575/volumes" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.137650 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" path="/var/lib/kubelet/pods/e70415f0-e1e4-4da7-a46a-ae3a01efa173/volumes" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.176945 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.235464 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.326790 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.344987 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.390341 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.464659 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.471470 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.472857 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.691205 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.717339 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.733814 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.860061 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 06:15:45 crc kubenswrapper[4685]: I1204 06:15:45.886046 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.007301 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.021732 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.044829 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.147021 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.249590 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.300774 4685 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.364324 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.582354 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.685373 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.743426 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 06:15:46 crc kubenswrapper[4685]: I1204 06:15:46.752127 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.172274 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.460053 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.503237 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.611189 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.644235 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.740788 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.799156 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.822623 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.894320 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.977853 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.984711 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 06:15:47 crc kubenswrapper[4685]: I1204 06:15:47.985169 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.086617 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.237476 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.478622 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.498191 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.590769 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.725174 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 06:15:48 crc kubenswrapper[4685]: I1204 06:15:48.881119 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.049624 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.593506 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.680798 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.743291 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.840192 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.850140 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 06:15:49 crc kubenswrapper[4685]: I1204 06:15:49.930467 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.032957 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.038892 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.280125 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.285438 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.300334 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.413845 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.461707 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.530901 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.567653 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.680756 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.680830 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.862919 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.863004 4685 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42" exitCode=137 Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.863071 4685 scope.go:117] "RemoveContainer" containerID="3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.863088 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874162 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874227 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874281 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874319 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874348 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874696 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874751 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874781 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.874817 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.889344 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.905644 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.976327 4685 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.976366 4685 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.976389 4685 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.976423 4685 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:50 crc kubenswrapper[4685]: I1204 06:15:50.976441 4685 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.005607 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.133477 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.202954 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.202966 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.416126 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.937176 4685 scope.go:117] "RemoveContainer" containerID="3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42" Dec 04 06:15:51 crc kubenswrapper[4685]: E1204 06:15:51.937783 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42\": container with ID starting with 3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42 not found: ID does not exist" containerID="3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.937808 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42"} err="failed to get container status \"3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42\": rpc error: code = NotFound desc = could not find container \"3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42\": container with ID starting with 3f776bcfbe4c92a58662bf79a41071ef6081192d914483a803f5e6b908ef7d42 not found: ID does not exist" Dec 04 06:15:51 crc kubenswrapper[4685]: I1204 06:15:51.937904 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 06:15:52 crc kubenswrapper[4685]: I1204 06:15:52.027124 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 06:15:52 crc kubenswrapper[4685]: I1204 06:15:52.348645 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 06:16:02 crc kubenswrapper[4685]: I1204 06:16:02.970490 4685 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923230 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49wxx"] Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.923881 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923898 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.923911 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923920 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.923930 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923938 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.923946 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923953 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.923966 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923973 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.923984 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.923991 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924002 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924009 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924022 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924028 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924038 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924045 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924057 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924066 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="extract-utilities" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924077 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924083 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="extract-content" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924092 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924098 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924108 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" containerName="installer" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924116 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" containerName="installer" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924125 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" containerName="marketplace-operator" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924132 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" containerName="marketplace-operator" Dec 04 06:16:11 crc kubenswrapper[4685]: E1204 06:16:11.924142 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924149 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924494 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a21c41-db1c-40c9-82d1-2fb0f33f4625" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924513 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2dcac27-38cb-4fd0-b9d3-54319c9d4b6e" containerName="installer" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924520 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2596408-3b90-481e-ad56-d5560edf0a96" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924530 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924540 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70415f0-e1e4-4da7-a46a-ae3a01efa173" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924546 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7736df92-b993-435a-ad74-4c2edd644ca3" containerName="marketplace-operator" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.924560 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2b22c7-e7ef-445f-a910-d068f1e17575" containerName="registry-server" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.925007 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.927122 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.927981 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.928353 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.929142 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.940338 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49wxx"] Dec 04 06:16:11 crc kubenswrapper[4685]: I1204 06:16:11.942322 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.000028 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh"] Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.000249 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" podUID="22360e53-a38b-4bc6-bb23-a993a434a880" containerName="route-controller-manager" containerID="cri-o://838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f" gracePeriod=30 Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.059333 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qkqs"] Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.060041 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" podUID="b50783dc-85ca-4dd8-9ad7-08224cf4934f" containerName="controller-manager" containerID="cri-o://785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b" gracePeriod=30 Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.061930 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c54a0f24-1600-4e6a-8c7c-00f003250e59-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.062005 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpghp\" (UniqueName: \"kubernetes.io/projected/c54a0f24-1600-4e6a-8c7c-00f003250e59-kube-api-access-dpghp\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.062043 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c54a0f24-1600-4e6a-8c7c-00f003250e59-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.163077 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpghp\" (UniqueName: \"kubernetes.io/projected/c54a0f24-1600-4e6a-8c7c-00f003250e59-kube-api-access-dpghp\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.163400 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c54a0f24-1600-4e6a-8c7c-00f003250e59-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.163591 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c54a0f24-1600-4e6a-8c7c-00f003250e59-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.164890 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c54a0f24-1600-4e6a-8c7c-00f003250e59-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.176625 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c54a0f24-1600-4e6a-8c7c-00f003250e59-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.190256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpghp\" (UniqueName: \"kubernetes.io/projected/c54a0f24-1600-4e6a-8c7c-00f003250e59-kube-api-access-dpghp\") pod \"marketplace-operator-79b997595-49wxx\" (UID: \"c54a0f24-1600-4e6a-8c7c-00f003250e59\") " pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.245381 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.456105 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.565675 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49wxx"] Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.573734 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.576974 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-config\") pod \"22360e53-a38b-4bc6-bb23-a993a434a880\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.577029 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-client-ca\") pod \"22360e53-a38b-4bc6-bb23-a993a434a880\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.577153 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gdgw\" (UniqueName: \"kubernetes.io/projected/22360e53-a38b-4bc6-bb23-a993a434a880-kube-api-access-2gdgw\") pod \"22360e53-a38b-4bc6-bb23-a993a434a880\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.577215 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22360e53-a38b-4bc6-bb23-a993a434a880-serving-cert\") pod \"22360e53-a38b-4bc6-bb23-a993a434a880\" (UID: \"22360e53-a38b-4bc6-bb23-a993a434a880\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.579357 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-client-ca" (OuterVolumeSpecName: "client-ca") pod "22360e53-a38b-4bc6-bb23-a993a434a880" (UID: "22360e53-a38b-4bc6-bb23-a993a434a880"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.579895 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-config" (OuterVolumeSpecName: "config") pod "22360e53-a38b-4bc6-bb23-a993a434a880" (UID: "22360e53-a38b-4bc6-bb23-a993a434a880"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.583999 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22360e53-a38b-4bc6-bb23-a993a434a880-kube-api-access-2gdgw" (OuterVolumeSpecName: "kube-api-access-2gdgw") pod "22360e53-a38b-4bc6-bb23-a993a434a880" (UID: "22360e53-a38b-4bc6-bb23-a993a434a880"). InnerVolumeSpecName "kube-api-access-2gdgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.584224 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22360e53-a38b-4bc6-bb23-a993a434a880-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "22360e53-a38b-4bc6-bb23-a993a434a880" (UID: "22360e53-a38b-4bc6-bb23-a993a434a880"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.677937 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-proxy-ca-bundles\") pod \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.678017 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b50783dc-85ca-4dd8-9ad7-08224cf4934f-serving-cert\") pod \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.678369 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-client-ca\") pod \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.678988 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-client-ca" (OuterVolumeSpecName: "client-ca") pod "b50783dc-85ca-4dd8-9ad7-08224cf4934f" (UID: "b50783dc-85ca-4dd8-9ad7-08224cf4934f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.679002 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b50783dc-85ca-4dd8-9ad7-08224cf4934f" (UID: "b50783dc-85ca-4dd8-9ad7-08224cf4934f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.679070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9cl2\" (UniqueName: \"kubernetes.io/projected/b50783dc-85ca-4dd8-9ad7-08224cf4934f-kube-api-access-k9cl2\") pod \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.689624 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b50783dc-85ca-4dd8-9ad7-08224cf4934f-kube-api-access-k9cl2" (OuterVolumeSpecName: "kube-api-access-k9cl2") pod "b50783dc-85ca-4dd8-9ad7-08224cf4934f" (UID: "b50783dc-85ca-4dd8-9ad7-08224cf4934f"). InnerVolumeSpecName "kube-api-access-k9cl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.689759 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-config\") pod \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\" (UID: \"b50783dc-85ca-4dd8-9ad7-08224cf4934f\") " Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690557 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690585 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gdgw\" (UniqueName: \"kubernetes.io/projected/22360e53-a38b-4bc6-bb23-a993a434a880-kube-api-access-2gdgw\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690601 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690620 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22360e53-a38b-4bc6-bb23-a993a434a880-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690636 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9cl2\" (UniqueName: \"kubernetes.io/projected/b50783dc-85ca-4dd8-9ad7-08224cf4934f-kube-api-access-k9cl2\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690649 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.690661 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/22360e53-a38b-4bc6-bb23-a993a434a880-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.691226 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-config" (OuterVolumeSpecName: "config") pod "b50783dc-85ca-4dd8-9ad7-08224cf4934f" (UID: "b50783dc-85ca-4dd8-9ad7-08224cf4934f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.692690 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b50783dc-85ca-4dd8-9ad7-08224cf4934f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b50783dc-85ca-4dd8-9ad7-08224cf4934f" (UID: "b50783dc-85ca-4dd8-9ad7-08224cf4934f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.791913 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b50783dc-85ca-4dd8-9ad7-08224cf4934f-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.791973 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b50783dc-85ca-4dd8-9ad7-08224cf4934f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.998207 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" event={"ID":"c54a0f24-1600-4e6a-8c7c-00f003250e59","Type":"ContainerStarted","Data":"1192314917f54c0d684d8d5903537003e44fe842ff9ed20e3f9bfe0345479be2"} Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.998581 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.998598 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" event={"ID":"c54a0f24-1600-4e6a-8c7c-00f003250e59","Type":"ContainerStarted","Data":"e3292acaf0fd53903c9f75f60dbf7ce33b7074487771c3c673232554961cc67d"} Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.999570 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-49wxx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" start-of-body= Dec 04 06:16:12 crc kubenswrapper[4685]: I1204 06:16:12.999607 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" podUID="c54a0f24-1600-4e6a-8c7c-00f003250e59" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.000537 4685 generic.go:334] "Generic (PLEG): container finished" podID="b50783dc-85ca-4dd8-9ad7-08224cf4934f" containerID="785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b" exitCode=0 Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.000585 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" event={"ID":"b50783dc-85ca-4dd8-9ad7-08224cf4934f","Type":"ContainerDied","Data":"785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b"} Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.000628 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.000640 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qkqs" event={"ID":"b50783dc-85ca-4dd8-9ad7-08224cf4934f","Type":"ContainerDied","Data":"e2d926136d27817814b36d6ec5f8a3b6dc40647513ebcaa248553a9245a2c127"} Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.000667 4685 scope.go:117] "RemoveContainer" containerID="785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.002783 4685 generic.go:334] "Generic (PLEG): container finished" podID="22360e53-a38b-4bc6-bb23-a993a434a880" containerID="838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f" exitCode=0 Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.002820 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" event={"ID":"22360e53-a38b-4bc6-bb23-a993a434a880","Type":"ContainerDied","Data":"838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f"} Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.002847 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" event={"ID":"22360e53-a38b-4bc6-bb23-a993a434a880","Type":"ContainerDied","Data":"e5fe358dde70a71aa1cf753d94fd2d232c0f044f29eedf37fa6060dae1a7928d"} Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.002993 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.017139 4685 scope.go:117] "RemoveContainer" containerID="785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b" Dec 04 06:16:13 crc kubenswrapper[4685]: E1204 06:16:13.017747 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b\": container with ID starting with 785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b not found: ID does not exist" containerID="785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.017846 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b"} err="failed to get container status \"785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b\": rpc error: code = NotFound desc = could not find container \"785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b\": container with ID starting with 785a943d5cbba62413f2c69efe82b43c7c8235d8a6a896cf8866d3859f6f173b not found: ID does not exist" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.017880 4685 scope.go:117] "RemoveContainer" containerID="838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.019658 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" podStartSLOduration=2.019640926 podStartE2EDuration="2.019640926s" podCreationTimestamp="2025-12-04 06:16:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:16:13.018575132 +0000 UTC m=+310.206805907" watchObservedRunningTime="2025-12-04 06:16:13.019640926 +0000 UTC m=+310.207871701" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.041157 4685 scope.go:117] "RemoveContainer" containerID="838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f" Dec 04 06:16:13 crc kubenswrapper[4685]: E1204 06:16:13.041938 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f\": container with ID starting with 838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f not found: ID does not exist" containerID="838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.041994 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f"} err="failed to get container status \"838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f\": rpc error: code = NotFound desc = could not find container \"838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f\": container with ID starting with 838805a992f0b89bd53b311fdbb0655e362e7ff60dfc392270d84add1325276f not found: ID does not exist" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.044568 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qkqs"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.047522 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qkqs"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.053647 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.059198 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b5sjh"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.133505 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22360e53-a38b-4bc6-bb23-a993a434a880" path="/var/lib/kubelet/pods/22360e53-a38b-4bc6-bb23-a993a434a880/volumes" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.134402 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b50783dc-85ca-4dd8-9ad7-08224cf4934f" path="/var/lib/kubelet/pods/b50783dc-85ca-4dd8-9ad7-08224cf4934f/volumes" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.645605 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-df6468bbf-wcdr9"] Dec 04 06:16:13 crc kubenswrapper[4685]: E1204 06:16:13.645962 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50783dc-85ca-4dd8-9ad7-08224cf4934f" containerName="controller-manager" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.645991 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50783dc-85ca-4dd8-9ad7-08224cf4934f" containerName="controller-manager" Dec 04 06:16:13 crc kubenswrapper[4685]: E1204 06:16:13.646021 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22360e53-a38b-4bc6-bb23-a993a434a880" containerName="route-controller-manager" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.646030 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="22360e53-a38b-4bc6-bb23-a993a434a880" containerName="route-controller-manager" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.646144 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b50783dc-85ca-4dd8-9ad7-08224cf4934f" containerName="controller-manager" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.646167 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="22360e53-a38b-4bc6-bb23-a993a434a880" containerName="route-controller-manager" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.646773 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.648304 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.648669 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.648817 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.649069 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.649256 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.649379 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.649868 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.650188 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.651959 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.652462 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.652719 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.652897 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.652948 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.652967 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.658558 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.664958 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-df6468bbf-wcdr9"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.669022 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql"] Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803546 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-proxy-ca-bundles\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803595 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bd89197-8478-4e8b-bd52-7c2081309e38-serving-cert\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803623 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-config\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803681 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm5xg\" (UniqueName: \"kubernetes.io/projected/3bd89197-8478-4e8b-bd52-7c2081309e38-kube-api-access-mm5xg\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803746 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9774d34d-7e88-4dbd-9893-5b8ff3e52867-serving-cert\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803773 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-client-ca\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803824 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-config\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803856 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-client-ca\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.803898 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h79r8\" (UniqueName: \"kubernetes.io/projected/9774d34d-7e88-4dbd-9893-5b8ff3e52867-kube-api-access-h79r8\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905120 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm5xg\" (UniqueName: \"kubernetes.io/projected/3bd89197-8478-4e8b-bd52-7c2081309e38-kube-api-access-mm5xg\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905205 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9774d34d-7e88-4dbd-9893-5b8ff3e52867-serving-cert\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905229 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-client-ca\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905254 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-config\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905276 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-client-ca\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h79r8\" (UniqueName: \"kubernetes.io/projected/9774d34d-7e88-4dbd-9893-5b8ff3e52867-kube-api-access-h79r8\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905326 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-proxy-ca-bundles\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905346 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bd89197-8478-4e8b-bd52-7c2081309e38-serving-cert\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.905361 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-config\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.906793 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-config\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.906798 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-client-ca\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.907054 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-client-ca\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.908963 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-config\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.909156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-proxy-ca-bundles\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.912851 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9774d34d-7e88-4dbd-9893-5b8ff3e52867-serving-cert\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.913178 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bd89197-8478-4e8b-bd52-7c2081309e38-serving-cert\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.927579 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm5xg\" (UniqueName: \"kubernetes.io/projected/3bd89197-8478-4e8b-bd52-7c2081309e38-kube-api-access-mm5xg\") pod \"route-controller-manager-75f94579cc-6s5ql\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.928725 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h79r8\" (UniqueName: \"kubernetes.io/projected/9774d34d-7e88-4dbd-9893-5b8ff3e52867-kube-api-access-h79r8\") pod \"controller-manager-df6468bbf-wcdr9\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.967552 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:13 crc kubenswrapper[4685]: I1204 06:16:13.982104 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:14 crc kubenswrapper[4685]: I1204 06:16:14.017973 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-49wxx" Dec 04 06:16:14 crc kubenswrapper[4685]: I1204 06:16:14.192554 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-df6468bbf-wcdr9"] Dec 04 06:16:14 crc kubenswrapper[4685]: I1204 06:16:14.270053 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql"] Dec 04 06:16:14 crc kubenswrapper[4685]: W1204 06:16:14.292418 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bd89197_8478_4e8b_bd52_7c2081309e38.slice/crio-0d5d911cb0bfb6c558ad5f832cbe0e1a542f90fd0cf487f901d5551e039f4252 WatchSource:0}: Error finding container 0d5d911cb0bfb6c558ad5f832cbe0e1a542f90fd0cf487f901d5551e039f4252: Status 404 returned error can't find the container with id 0d5d911cb0bfb6c558ad5f832cbe0e1a542f90fd0cf487f901d5551e039f4252 Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.021654 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.023798 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.023856 4685 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1d5cac1482b7d8a618a67d1affcc35dda0d5481c7a627db3d606d899a6bd9bb0" exitCode=137 Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.023929 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1d5cac1482b7d8a618a67d1affcc35dda0d5481c7a627db3d606d899a6bd9bb0"} Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.023956 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"58576a573fca0f8796186faab5615c77c02cb2195e55be9e5e87273147aec8fd"} Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.023972 4685 scope.go:117] "RemoveContainer" containerID="801bd436e2fbf83b5768802af50628b458b9cefdcd3cd987c7557e3612794de3" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.026303 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" event={"ID":"9774d34d-7e88-4dbd-9893-5b8ff3e52867","Type":"ContainerStarted","Data":"f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e"} Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.026343 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" event={"ID":"9774d34d-7e88-4dbd-9893-5b8ff3e52867","Type":"ContainerStarted","Data":"40ff5b4599ffb3d7d876e354daed67196efe8426dc8d04427280506d81c446c0"} Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.026821 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.028998 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" event={"ID":"3bd89197-8478-4e8b-bd52-7c2081309e38","Type":"ContainerStarted","Data":"ad34c6d0f2cc7027277ab14c1fdd4843237dd1b1b740bd79f6d86407a27d1197"} Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.029029 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" event={"ID":"3bd89197-8478-4e8b-bd52-7c2081309e38","Type":"ContainerStarted","Data":"0d5d911cb0bfb6c558ad5f832cbe0e1a542f90fd0cf487f901d5551e039f4252"} Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.029118 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.033880 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.034447 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.057466 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" podStartSLOduration=3.057448713 podStartE2EDuration="3.057448713s" podCreationTimestamp="2025-12-04 06:16:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:16:15.057115033 +0000 UTC m=+312.245345818" watchObservedRunningTime="2025-12-04 06:16:15.057448713 +0000 UTC m=+312.245679498" Dec 04 06:16:15 crc kubenswrapper[4685]: I1204 06:16:15.076994 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" podStartSLOduration=3.076976687 podStartE2EDuration="3.076976687s" podCreationTimestamp="2025-12-04 06:16:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:16:15.075901432 +0000 UTC m=+312.264132207" watchObservedRunningTime="2025-12-04 06:16:15.076976687 +0000 UTC m=+312.265207462" Dec 04 06:16:16 crc kubenswrapper[4685]: I1204 06:16:16.038016 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 04 06:16:21 crc kubenswrapper[4685]: I1204 06:16:21.929734 4685 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podf85e55b1a89d02b0cb034b1ea31ed45a"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podf85e55b1a89d02b0cb034b1ea31ed45a] : Timed out while waiting for systemd to remove kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice" Dec 04 06:16:23 crc kubenswrapper[4685]: I1204 06:16:23.573481 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:16:24 crc kubenswrapper[4685]: I1204 06:16:24.111680 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:16:24 crc kubenswrapper[4685]: I1204 06:16:24.115646 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.074851 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g2sbc"] Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.076457 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.079101 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.085585 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2sbc"] Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.102802 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.158763 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk6ck\" (UniqueName: \"kubernetes.io/projected/492c20de-7b85-47aa-b498-8531fc23bf24-kube-api-access-qk6ck\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.159181 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492c20de-7b85-47aa-b498-8531fc23bf24-catalog-content\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.159399 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492c20de-7b85-47aa-b498-8531fc23bf24-utilities\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.261551 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk6ck\" (UniqueName: \"kubernetes.io/projected/492c20de-7b85-47aa-b498-8531fc23bf24-kube-api-access-qk6ck\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.262209 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492c20de-7b85-47aa-b498-8531fc23bf24-catalog-content\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.262549 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492c20de-7b85-47aa-b498-8531fc23bf24-utilities\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.262703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492c20de-7b85-47aa-b498-8531fc23bf24-catalog-content\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.263197 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492c20de-7b85-47aa-b498-8531fc23bf24-utilities\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.275027 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rlc5g"] Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.276144 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.279282 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.291380 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rlc5g"] Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.298373 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk6ck\" (UniqueName: \"kubernetes.io/projected/492c20de-7b85-47aa-b498-8531fc23bf24-kube-api-access-qk6ck\") pod \"redhat-marketplace-g2sbc\" (UID: \"492c20de-7b85-47aa-b498-8531fc23bf24\") " pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.364911 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3085a-8ad1-45e0-8df7-f43cdb14da32-catalog-content\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.365443 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3085a-8ad1-45e0-8df7-f43cdb14da32-utilities\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.365476 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khmxz\" (UniqueName: \"kubernetes.io/projected/61f3085a-8ad1-45e0-8df7-f43cdb14da32-kube-api-access-khmxz\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.413968 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.466615 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3085a-8ad1-45e0-8df7-f43cdb14da32-catalog-content\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.466667 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3085a-8ad1-45e0-8df7-f43cdb14da32-utilities\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.466704 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khmxz\" (UniqueName: \"kubernetes.io/projected/61f3085a-8ad1-45e0-8df7-f43cdb14da32-kube-api-access-khmxz\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.467330 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3085a-8ad1-45e0-8df7-f43cdb14da32-utilities\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.467400 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3085a-8ad1-45e0-8df7-f43cdb14da32-catalog-content\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.491253 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khmxz\" (UniqueName: \"kubernetes.io/projected/61f3085a-8ad1-45e0-8df7-f43cdb14da32-kube-api-access-khmxz\") pod \"redhat-operators-rlc5g\" (UID: \"61f3085a-8ad1-45e0-8df7-f43cdb14da32\") " pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.638318 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:25 crc kubenswrapper[4685]: I1204 06:16:25.862609 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2sbc"] Dec 04 06:16:25 crc kubenswrapper[4685]: W1204 06:16:25.865238 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod492c20de_7b85_47aa_b498_8531fc23bf24.slice/crio-1694cb9e90cdc2504c7cdf3b4a058886c5e6d6ddc4000149173ae72e32dc041d WatchSource:0}: Error finding container 1694cb9e90cdc2504c7cdf3b4a058886c5e6d6ddc4000149173ae72e32dc041d: Status 404 returned error can't find the container with id 1694cb9e90cdc2504c7cdf3b4a058886c5e6d6ddc4000149173ae72e32dc041d Dec 04 06:16:26 crc kubenswrapper[4685]: I1204 06:16:26.090765 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rlc5g"] Dec 04 06:16:26 crc kubenswrapper[4685]: W1204 06:16:26.094226 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61f3085a_8ad1_45e0_8df7_f43cdb14da32.slice/crio-20e368cb5ebcd3d701df6a9f891c5ec9f201b5bfa14a83e65889bed4e3f899d0 WatchSource:0}: Error finding container 20e368cb5ebcd3d701df6a9f891c5ec9f201b5bfa14a83e65889bed4e3f899d0: Status 404 returned error can't find the container with id 20e368cb5ebcd3d701df6a9f891c5ec9f201b5bfa14a83e65889bed4e3f899d0 Dec 04 06:16:26 crc kubenswrapper[4685]: I1204 06:16:26.101896 4685 generic.go:334] "Generic (PLEG): container finished" podID="492c20de-7b85-47aa-b498-8531fc23bf24" containerID="9df8f0008fdf488ca76c277d5fe51a7644af02e3246e1b20ec2691c8ac0d4926" exitCode=0 Dec 04 06:16:26 crc kubenswrapper[4685]: I1204 06:16:26.101980 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2sbc" event={"ID":"492c20de-7b85-47aa-b498-8531fc23bf24","Type":"ContainerDied","Data":"9df8f0008fdf488ca76c277d5fe51a7644af02e3246e1b20ec2691c8ac0d4926"} Dec 04 06:16:26 crc kubenswrapper[4685]: I1204 06:16:26.102051 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2sbc" event={"ID":"492c20de-7b85-47aa-b498-8531fc23bf24","Type":"ContainerStarted","Data":"1694cb9e90cdc2504c7cdf3b4a058886c5e6d6ddc4000149173ae72e32dc041d"} Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.109911 4685 generic.go:334] "Generic (PLEG): container finished" podID="492c20de-7b85-47aa-b498-8531fc23bf24" containerID="080422c6e0205f0b30dcec6bb30a15c95134fa76bce14e84763cec2f696b0ec4" exitCode=0 Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.109983 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2sbc" event={"ID":"492c20de-7b85-47aa-b498-8531fc23bf24","Type":"ContainerDied","Data":"080422c6e0205f0b30dcec6bb30a15c95134fa76bce14e84763cec2f696b0ec4"} Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.114154 4685 generic.go:334] "Generic (PLEG): container finished" podID="61f3085a-8ad1-45e0-8df7-f43cdb14da32" containerID="7b7f127746a8d02e31582163cfe728147bf8f50fdf0161b759cb80ea08aa88c5" exitCode=0 Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.114208 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlc5g" event={"ID":"61f3085a-8ad1-45e0-8df7-f43cdb14da32","Type":"ContainerDied","Data":"7b7f127746a8d02e31582163cfe728147bf8f50fdf0161b759cb80ea08aa88c5"} Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.114249 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlc5g" event={"ID":"61f3085a-8ad1-45e0-8df7-f43cdb14da32","Type":"ContainerStarted","Data":"20e368cb5ebcd3d701df6a9f891c5ec9f201b5bfa14a83e65889bed4e3f899d0"} Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.672527 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8bjlq"] Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.674240 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.676930 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.685173 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8bjlq"] Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.704320 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a563d518-9568-45f7-8cf2-69c0b2fe9835-catalog-content\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.704495 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a563d518-9568-45f7-8cf2-69c0b2fe9835-utilities\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.704799 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clz2h\" (UniqueName: \"kubernetes.io/projected/a563d518-9568-45f7-8cf2-69c0b2fe9835-kube-api-access-clz2h\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.806191 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a563d518-9568-45f7-8cf2-69c0b2fe9835-catalog-content\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.806248 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a563d518-9568-45f7-8cf2-69c0b2fe9835-utilities\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.806288 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clz2h\" (UniqueName: \"kubernetes.io/projected/a563d518-9568-45f7-8cf2-69c0b2fe9835-kube-api-access-clz2h\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.806978 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a563d518-9568-45f7-8cf2-69c0b2fe9835-utilities\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.806990 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a563d518-9568-45f7-8cf2-69c0b2fe9835-catalog-content\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.840499 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clz2h\" (UniqueName: \"kubernetes.io/projected/a563d518-9568-45f7-8cf2-69c0b2fe9835-kube-api-access-clz2h\") pod \"certified-operators-8bjlq\" (UID: \"a563d518-9568-45f7-8cf2-69c0b2fe9835\") " pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.869795 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-brpm6"] Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.870802 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.873011 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.888852 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brpm6"] Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.908131 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddbpc\" (UniqueName: \"kubernetes.io/projected/50ea99f6-03dd-4c92-90d8-1d57fe44582f-kube-api-access-ddbpc\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.908205 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea99f6-03dd-4c92-90d8-1d57fe44582f-utilities\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:27 crc kubenswrapper[4685]: I1204 06:16:27.908299 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea99f6-03dd-4c92-90d8-1d57fe44582f-catalog-content\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.005443 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.009895 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddbpc\" (UniqueName: \"kubernetes.io/projected/50ea99f6-03dd-4c92-90d8-1d57fe44582f-kube-api-access-ddbpc\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.009953 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea99f6-03dd-4c92-90d8-1d57fe44582f-utilities\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.009991 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea99f6-03dd-4c92-90d8-1d57fe44582f-catalog-content\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.010552 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ea99f6-03dd-4c92-90d8-1d57fe44582f-catalog-content\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.010581 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ea99f6-03dd-4c92-90d8-1d57fe44582f-utilities\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.036459 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddbpc\" (UniqueName: \"kubernetes.io/projected/50ea99f6-03dd-4c92-90d8-1d57fe44582f-kube-api-access-ddbpc\") pod \"community-operators-brpm6\" (UID: \"50ea99f6-03dd-4c92-90d8-1d57fe44582f\") " pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.135638 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2sbc" event={"ID":"492c20de-7b85-47aa-b498-8531fc23bf24","Type":"ContainerStarted","Data":"2ec3304dc0966bba1bdd240ec375b8f5d32dd674c551621c2bcfc7e63c8259f0"} Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.144019 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlc5g" event={"ID":"61f3085a-8ad1-45e0-8df7-f43cdb14da32","Type":"ContainerStarted","Data":"c7e419f68a9e3427a407b36e598766a0923d84e7cb5b1c5830b3be5a0214cffa"} Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.157439 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g2sbc" podStartSLOduration=1.7807785809999999 podStartE2EDuration="3.157394472s" podCreationTimestamp="2025-12-04 06:16:25 +0000 UTC" firstStartedPulling="2025-12-04 06:16:26.110577769 +0000 UTC m=+323.298808544" lastFinishedPulling="2025-12-04 06:16:27.48719362 +0000 UTC m=+324.675424435" observedRunningTime="2025-12-04 06:16:28.15277427 +0000 UTC m=+325.341005055" watchObservedRunningTime="2025-12-04 06:16:28.157394472 +0000 UTC m=+325.345625257" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.193765 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.451901 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8bjlq"] Dec 04 06:16:28 crc kubenswrapper[4685]: W1204 06:16:28.470482 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda563d518_9568_45f7_8cf2_69c0b2fe9835.slice/crio-330a9257f0900e4b25949fc271929758e97b0e22877dfe38b596f9bf336410d1 WatchSource:0}: Error finding container 330a9257f0900e4b25949fc271929758e97b0e22877dfe38b596f9bf336410d1: Status 404 returned error can't find the container with id 330a9257f0900e4b25949fc271929758e97b0e22877dfe38b596f9bf336410d1 Dec 04 06:16:28 crc kubenswrapper[4685]: I1204 06:16:28.662380 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brpm6"] Dec 04 06:16:28 crc kubenswrapper[4685]: W1204 06:16:28.685735 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50ea99f6_03dd_4c92_90d8_1d57fe44582f.slice/crio-e0ef7423f4efc44ee971f3c46e1f5993b6cb28d928861699d0e4a29aae36608c WatchSource:0}: Error finding container e0ef7423f4efc44ee971f3c46e1f5993b6cb28d928861699d0e4a29aae36608c: Status 404 returned error can't find the container with id e0ef7423f4efc44ee971f3c46e1f5993b6cb28d928861699d0e4a29aae36608c Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.151486 4685 generic.go:334] "Generic (PLEG): container finished" podID="61f3085a-8ad1-45e0-8df7-f43cdb14da32" containerID="c7e419f68a9e3427a407b36e598766a0923d84e7cb5b1c5830b3be5a0214cffa" exitCode=0 Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.151608 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlc5g" event={"ID":"61f3085a-8ad1-45e0-8df7-f43cdb14da32","Type":"ContainerDied","Data":"c7e419f68a9e3427a407b36e598766a0923d84e7cb5b1c5830b3be5a0214cffa"} Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.155692 4685 generic.go:334] "Generic (PLEG): container finished" podID="50ea99f6-03dd-4c92-90d8-1d57fe44582f" containerID="43b73c05341deb480fdb5a8cf6cbaddd0301294a7f9e90e2201962bfed63801b" exitCode=0 Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.155752 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brpm6" event={"ID":"50ea99f6-03dd-4c92-90d8-1d57fe44582f","Type":"ContainerDied","Data":"43b73c05341deb480fdb5a8cf6cbaddd0301294a7f9e90e2201962bfed63801b"} Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.155826 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brpm6" event={"ID":"50ea99f6-03dd-4c92-90d8-1d57fe44582f","Type":"ContainerStarted","Data":"e0ef7423f4efc44ee971f3c46e1f5993b6cb28d928861699d0e4a29aae36608c"} Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.158388 4685 generic.go:334] "Generic (PLEG): container finished" podID="a563d518-9568-45f7-8cf2-69c0b2fe9835" containerID="19d8752a572d7bcb2b07326819c8b6fb527b4f10c1910a5c071eb6bc68a1a821" exitCode=0 Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.159147 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bjlq" event={"ID":"a563d518-9568-45f7-8cf2-69c0b2fe9835","Type":"ContainerDied","Data":"19d8752a572d7bcb2b07326819c8b6fb527b4f10c1910a5c071eb6bc68a1a821"} Dec 04 06:16:29 crc kubenswrapper[4685]: I1204 06:16:29.159202 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bjlq" event={"ID":"a563d518-9568-45f7-8cf2-69c0b2fe9835","Type":"ContainerStarted","Data":"330a9257f0900e4b25949fc271929758e97b0e22877dfe38b596f9bf336410d1"} Dec 04 06:16:30 crc kubenswrapper[4685]: I1204 06:16:30.163841 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bjlq" event={"ID":"a563d518-9568-45f7-8cf2-69c0b2fe9835","Type":"ContainerStarted","Data":"d38fca9aa08be5f131b11bdbdb3dee90f5fb32540f855e9cfcdb7e2957bf2d83"} Dec 04 06:16:30 crc kubenswrapper[4685]: I1204 06:16:30.166464 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlc5g" event={"ID":"61f3085a-8ad1-45e0-8df7-f43cdb14da32","Type":"ContainerStarted","Data":"2c2074d579521d10455bb4c2b5480cc117e7ef080a26f583bda1e272ea7e99f8"} Dec 04 06:16:30 crc kubenswrapper[4685]: I1204 06:16:30.168557 4685 generic.go:334] "Generic (PLEG): container finished" podID="50ea99f6-03dd-4c92-90d8-1d57fe44582f" containerID="f021c34a32f2439af537ec61a1b38fe9b8989f240ac56d3963c55c73ea5f9e06" exitCode=0 Dec 04 06:16:30 crc kubenswrapper[4685]: I1204 06:16:30.169057 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brpm6" event={"ID":"50ea99f6-03dd-4c92-90d8-1d57fe44582f","Type":"ContainerDied","Data":"f021c34a32f2439af537ec61a1b38fe9b8989f240ac56d3963c55c73ea5f9e06"} Dec 04 06:16:30 crc kubenswrapper[4685]: I1204 06:16:30.222657 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rlc5g" podStartSLOduration=2.552103377 podStartE2EDuration="5.222642513s" podCreationTimestamp="2025-12-04 06:16:25 +0000 UTC" firstStartedPulling="2025-12-04 06:16:27.115341836 +0000 UTC m=+324.303572611" lastFinishedPulling="2025-12-04 06:16:29.785880952 +0000 UTC m=+326.974111747" observedRunningTime="2025-12-04 06:16:30.217935158 +0000 UTC m=+327.406165933" watchObservedRunningTime="2025-12-04 06:16:30.222642513 +0000 UTC m=+327.410873288" Dec 04 06:16:31 crc kubenswrapper[4685]: I1204 06:16:31.176050 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brpm6" event={"ID":"50ea99f6-03dd-4c92-90d8-1d57fe44582f","Type":"ContainerStarted","Data":"843bc39ab17d65490b7b76bc392b23315590fcab4dd54c4d18ff24d01b2ac8c5"} Dec 04 06:16:31 crc kubenswrapper[4685]: I1204 06:16:31.178973 4685 generic.go:334] "Generic (PLEG): container finished" podID="a563d518-9568-45f7-8cf2-69c0b2fe9835" containerID="d38fca9aa08be5f131b11bdbdb3dee90f5fb32540f855e9cfcdb7e2957bf2d83" exitCode=0 Dec 04 06:16:31 crc kubenswrapper[4685]: I1204 06:16:31.179072 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bjlq" event={"ID":"a563d518-9568-45f7-8cf2-69c0b2fe9835","Type":"ContainerDied","Data":"d38fca9aa08be5f131b11bdbdb3dee90f5fb32540f855e9cfcdb7e2957bf2d83"} Dec 04 06:16:31 crc kubenswrapper[4685]: I1204 06:16:31.201193 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-brpm6" podStartSLOduration=2.800527655 podStartE2EDuration="4.20116299s" podCreationTimestamp="2025-12-04 06:16:27 +0000 UTC" firstStartedPulling="2025-12-04 06:16:29.157210087 +0000 UTC m=+326.345440862" lastFinishedPulling="2025-12-04 06:16:30.557845422 +0000 UTC m=+327.746076197" observedRunningTime="2025-12-04 06:16:31.195635869 +0000 UTC m=+328.383866664" watchObservedRunningTime="2025-12-04 06:16:31.20116299 +0000 UTC m=+328.389393765" Dec 04 06:16:32 crc kubenswrapper[4685]: I1204 06:16:32.188382 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bjlq" event={"ID":"a563d518-9568-45f7-8cf2-69c0b2fe9835","Type":"ContainerStarted","Data":"9a4f94003550971cef1e2ca848508d665cf035e885cce4a7038abfaa5ffc1794"} Dec 04 06:16:32 crc kubenswrapper[4685]: I1204 06:16:32.208935 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8bjlq" podStartSLOduration=2.7182079420000003 podStartE2EDuration="5.208908743s" podCreationTimestamp="2025-12-04 06:16:27 +0000 UTC" firstStartedPulling="2025-12-04 06:16:29.159968638 +0000 UTC m=+326.348199413" lastFinishedPulling="2025-12-04 06:16:31.650669439 +0000 UTC m=+328.838900214" observedRunningTime="2025-12-04 06:16:32.20736285 +0000 UTC m=+329.395593635" watchObservedRunningTime="2025-12-04 06:16:32.208908743 +0000 UTC m=+329.397139528" Dec 04 06:16:34 crc kubenswrapper[4685]: I1204 06:16:34.380473 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql"] Dec 04 06:16:34 crc kubenswrapper[4685]: I1204 06:16:34.381316 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" podUID="3bd89197-8478-4e8b-bd52-7c2081309e38" containerName="route-controller-manager" containerID="cri-o://ad34c6d0f2cc7027277ab14c1fdd4843237dd1b1b740bd79f6d86407a27d1197" gracePeriod=30 Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.208320 4685 generic.go:334] "Generic (PLEG): container finished" podID="3bd89197-8478-4e8b-bd52-7c2081309e38" containerID="ad34c6d0f2cc7027277ab14c1fdd4843237dd1b1b740bd79f6d86407a27d1197" exitCode=0 Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.208474 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" event={"ID":"3bd89197-8478-4e8b-bd52-7c2081309e38","Type":"ContainerDied","Data":"ad34c6d0f2cc7027277ab14c1fdd4843237dd1b1b740bd79f6d86407a27d1197"} Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.415044 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.415135 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.469309 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.638741 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.639103 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:35 crc kubenswrapper[4685]: I1204 06:16:35.688794 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.145082 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.170807 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w"] Dec 04 06:16:36 crc kubenswrapper[4685]: E1204 06:16:36.171051 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd89197-8478-4e8b-bd52-7c2081309e38" containerName="route-controller-manager" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.171066 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd89197-8478-4e8b-bd52-7c2081309e38" containerName="route-controller-manager" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.171190 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bd89197-8478-4e8b-bd52-7c2081309e38" containerName="route-controller-manager" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.171653 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.216324 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.216683 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql" event={"ID":"3bd89197-8478-4e8b-bd52-7c2081309e38","Type":"ContainerDied","Data":"0d5d911cb0bfb6c558ad5f832cbe0e1a542f90fd0cf487f901d5551e039f4252"} Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.216759 4685 scope.go:117] "RemoveContainer" containerID="ad34c6d0f2cc7027277ab14c1fdd4843237dd1b1b740bd79f6d86407a27d1197" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.225552 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w"] Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.230000 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bd89197-8478-4e8b-bd52-7c2081309e38-serving-cert\") pod \"3bd89197-8478-4e8b-bd52-7c2081309e38\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.230052 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-client-ca\") pod \"3bd89197-8478-4e8b-bd52-7c2081309e38\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.230112 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-config\") pod \"3bd89197-8478-4e8b-bd52-7c2081309e38\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.230211 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm5xg\" (UniqueName: \"kubernetes.io/projected/3bd89197-8478-4e8b-bd52-7c2081309e38-kube-api-access-mm5xg\") pod \"3bd89197-8478-4e8b-bd52-7c2081309e38\" (UID: \"3bd89197-8478-4e8b-bd52-7c2081309e38\") " Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.232443 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-client-ca" (OuterVolumeSpecName: "client-ca") pod "3bd89197-8478-4e8b-bd52-7c2081309e38" (UID: "3bd89197-8478-4e8b-bd52-7c2081309e38"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.233012 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-config" (OuterVolumeSpecName: "config") pod "3bd89197-8478-4e8b-bd52-7c2081309e38" (UID: "3bd89197-8478-4e8b-bd52-7c2081309e38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.238884 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bd89197-8478-4e8b-bd52-7c2081309e38-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3bd89197-8478-4e8b-bd52-7c2081309e38" (UID: "3bd89197-8478-4e8b-bd52-7c2081309e38"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.245607 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bd89197-8478-4e8b-bd52-7c2081309e38-kube-api-access-mm5xg" (OuterVolumeSpecName: "kube-api-access-mm5xg") pod "3bd89197-8478-4e8b-bd52-7c2081309e38" (UID: "3bd89197-8478-4e8b-bd52-7c2081309e38"). InnerVolumeSpecName "kube-api-access-mm5xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.246130 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2nf7t"] Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.246804 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.260843 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2nf7t"] Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.282527 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g2sbc" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.307481 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rlc5g" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.331955 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-registry-tls\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332050 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5788j\" (UniqueName: \"kubernetes.io/projected/e0bba824-52ab-4951-a50a-b1cd257984bc-kube-api-access-5788j\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332091 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvzkw\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-kube-api-access-nvzkw\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332115 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dc689283-c531-45ab-b07b-39a39c74872d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332135 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc689283-c531-45ab-b07b-39a39c74872d-trusted-ca\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332159 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0bba824-52ab-4951-a50a-b1cd257984bc-serving-cert\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332182 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0bba824-52ab-4951-a50a-b1cd257984bc-client-ca\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332210 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-bound-sa-token\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332244 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dc689283-c531-45ab-b07b-39a39c74872d-registry-certificates\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332278 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0bba824-52ab-4951-a50a-b1cd257984bc-config\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332332 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dc689283-c531-45ab-b07b-39a39c74872d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332383 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332400 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm5xg\" (UniqueName: \"kubernetes.io/projected/3bd89197-8478-4e8b-bd52-7c2081309e38-kube-api-access-mm5xg\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332427 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bd89197-8478-4e8b-bd52-7c2081309e38-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.332437 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bd89197-8478-4e8b-bd52-7c2081309e38-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.377459 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.433943 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0bba824-52ab-4951-a50a-b1cd257984bc-config\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434029 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dc689283-c531-45ab-b07b-39a39c74872d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434067 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-registry-tls\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434107 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5788j\" (UniqueName: \"kubernetes.io/projected/e0bba824-52ab-4951-a50a-b1cd257984bc-kube-api-access-5788j\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434135 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvzkw\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-kube-api-access-nvzkw\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434166 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dc689283-c531-45ab-b07b-39a39c74872d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434189 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc689283-c531-45ab-b07b-39a39c74872d-trusted-ca\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434209 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0bba824-52ab-4951-a50a-b1cd257984bc-serving-cert\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434233 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0bba824-52ab-4951-a50a-b1cd257984bc-client-ca\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434312 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-bound-sa-token\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.434344 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dc689283-c531-45ab-b07b-39a39c74872d-registry-certificates\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.435781 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dc689283-c531-45ab-b07b-39a39c74872d-registry-certificates\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.436183 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0bba824-52ab-4951-a50a-b1cd257984bc-config\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.436202 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc689283-c531-45ab-b07b-39a39c74872d-trusted-ca\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.436453 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dc689283-c531-45ab-b07b-39a39c74872d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.436595 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0bba824-52ab-4951-a50a-b1cd257984bc-client-ca\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.441867 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-registry-tls\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.444486 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dc689283-c531-45ab-b07b-39a39c74872d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.444682 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0bba824-52ab-4951-a50a-b1cd257984bc-serving-cert\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.450635 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvzkw\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-kube-api-access-nvzkw\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.451775 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc689283-c531-45ab-b07b-39a39c74872d-bound-sa-token\") pod \"image-registry-66df7c8f76-2nf7t\" (UID: \"dc689283-c531-45ab-b07b-39a39c74872d\") " pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.456000 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5788j\" (UniqueName: \"kubernetes.io/projected/e0bba824-52ab-4951-a50a-b1cd257984bc-kube-api-access-5788j\") pod \"route-controller-manager-7764d4fd9-kft8w\" (UID: \"e0bba824-52ab-4951-a50a-b1cd257984bc\") " pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.497133 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.549461 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql"] Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.552850 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75f94579cc-6s5ql"] Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.593360 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:36 crc kubenswrapper[4685]: I1204 06:16:36.891676 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w"] Dec 04 06:16:37 crc kubenswrapper[4685]: I1204 06:16:37.006642 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2nf7t"] Dec 04 06:16:37 crc kubenswrapper[4685]: W1204 06:16:37.010642 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc689283_c531_45ab_b07b_39a39c74872d.slice/crio-6df7c11f9d77952c1ccc474bda670229be7fa27f8f9b74e234f798e394df2f8a WatchSource:0}: Error finding container 6df7c11f9d77952c1ccc474bda670229be7fa27f8f9b74e234f798e394df2f8a: Status 404 returned error can't find the container with id 6df7c11f9d77952c1ccc474bda670229be7fa27f8f9b74e234f798e394df2f8a Dec 04 06:16:37 crc kubenswrapper[4685]: I1204 06:16:37.133525 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bd89197-8478-4e8b-bd52-7c2081309e38" path="/var/lib/kubelet/pods/3bd89197-8478-4e8b-bd52-7c2081309e38/volumes" Dec 04 06:16:37 crc kubenswrapper[4685]: I1204 06:16:37.225755 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" event={"ID":"dc689283-c531-45ab-b07b-39a39c74872d","Type":"ContainerStarted","Data":"6df7c11f9d77952c1ccc474bda670229be7fa27f8f9b74e234f798e394df2f8a"} Dec 04 06:16:37 crc kubenswrapper[4685]: I1204 06:16:37.227256 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" event={"ID":"e0bba824-52ab-4951-a50a-b1cd257984bc","Type":"ContainerStarted","Data":"501a959c21b95d6ef93c7f6b4c8791e163fab7bb21d4618b5b507727073078f9"} Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.005846 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.006263 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.052771 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.194874 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.194945 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.234509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" event={"ID":"dc689283-c531-45ab-b07b-39a39c74872d","Type":"ContainerStarted","Data":"1d6d15dae44dcd9652a2a2ea00cd59a283e233be03b25c598519f9b05191b288"} Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.234656 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.236023 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" event={"ID":"e0bba824-52ab-4951-a50a-b1cd257984bc","Type":"ContainerStarted","Data":"44bd0bbebc9bd927f72aaf885079a15a0c5c3b127e7adb9c0f26d146c9b2f6ab"} Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.236673 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.241689 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.260614 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" podStartSLOduration=2.260585679 podStartE2EDuration="2.260585679s" podCreationTimestamp="2025-12-04 06:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:16:38.256362073 +0000 UTC m=+335.444592848" watchObservedRunningTime="2025-12-04 06:16:38.260585679 +0000 UTC m=+335.448816454" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.265870 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.294866 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7764d4fd9-kft8w" podStartSLOduration=4.294843755 podStartE2EDuration="4.294843755s" podCreationTimestamp="2025-12-04 06:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:16:38.28977243 +0000 UTC m=+335.478003225" watchObservedRunningTime="2025-12-04 06:16:38.294843755 +0000 UTC m=+335.483074530" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.301217 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8bjlq" Dec 04 06:16:38 crc kubenswrapper[4685]: I1204 06:16:38.334957 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-brpm6" Dec 04 06:16:50 crc kubenswrapper[4685]: I1204 06:16:50.828044 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-df6468bbf-wcdr9"] Dec 04 06:16:50 crc kubenswrapper[4685]: I1204 06:16:50.828851 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" podUID="9774d34d-7e88-4dbd-9893-5b8ff3e52867" containerName="controller-manager" containerID="cri-o://f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e" gracePeriod=30 Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.206481 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.323278 4685 generic.go:334] "Generic (PLEG): container finished" podID="9774d34d-7e88-4dbd-9893-5b8ff3e52867" containerID="f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e" exitCode=0 Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.323347 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" event={"ID":"9774d34d-7e88-4dbd-9893-5b8ff3e52867","Type":"ContainerDied","Data":"f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e"} Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.323389 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" event={"ID":"9774d34d-7e88-4dbd-9893-5b8ff3e52867","Type":"ContainerDied","Data":"40ff5b4599ffb3d7d876e354daed67196efe8426dc8d04427280506d81c446c0"} Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.323443 4685 scope.go:117] "RemoveContainer" containerID="f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.323604 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-df6468bbf-wcdr9" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.347079 4685 scope.go:117] "RemoveContainer" containerID="f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e" Dec 04 06:16:51 crc kubenswrapper[4685]: E1204 06:16:51.347607 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e\": container with ID starting with f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e not found: ID does not exist" containerID="f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.347691 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e"} err="failed to get container status \"f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e\": rpc error: code = NotFound desc = could not find container \"f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e\": container with ID starting with f99d381116bf33036bdd9308eb1e867bc05bd2a115609c284a37bc9f6913b60e not found: ID does not exist" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.351221 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-client-ca\") pod \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.351338 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-proxy-ca-bundles\") pod \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.351453 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-config\") pod \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.351541 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h79r8\" (UniqueName: \"kubernetes.io/projected/9774d34d-7e88-4dbd-9893-5b8ff3e52867-kube-api-access-h79r8\") pod \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.351624 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9774d34d-7e88-4dbd-9893-5b8ff3e52867-serving-cert\") pod \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\" (UID: \"9774d34d-7e88-4dbd-9893-5b8ff3e52867\") " Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.352483 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-client-ca" (OuterVolumeSpecName: "client-ca") pod "9774d34d-7e88-4dbd-9893-5b8ff3e52867" (UID: "9774d34d-7e88-4dbd-9893-5b8ff3e52867"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.352541 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9774d34d-7e88-4dbd-9893-5b8ff3e52867" (UID: "9774d34d-7e88-4dbd-9893-5b8ff3e52867"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.353066 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-config" (OuterVolumeSpecName: "config") pod "9774d34d-7e88-4dbd-9893-5b8ff3e52867" (UID: "9774d34d-7e88-4dbd-9893-5b8ff3e52867"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.360711 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9774d34d-7e88-4dbd-9893-5b8ff3e52867-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9774d34d-7e88-4dbd-9893-5b8ff3e52867" (UID: "9774d34d-7e88-4dbd-9893-5b8ff3e52867"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.360710 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9774d34d-7e88-4dbd-9893-5b8ff3e52867-kube-api-access-h79r8" (OuterVolumeSpecName: "kube-api-access-h79r8") pod "9774d34d-7e88-4dbd-9893-5b8ff3e52867" (UID: "9774d34d-7e88-4dbd-9893-5b8ff3e52867"). InnerVolumeSpecName "kube-api-access-h79r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.453099 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.453181 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.453208 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774d34d-7e88-4dbd-9893-5b8ff3e52867-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.453227 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h79r8\" (UniqueName: \"kubernetes.io/projected/9774d34d-7e88-4dbd-9893-5b8ff3e52867-kube-api-access-h79r8\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.453247 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9774d34d-7e88-4dbd-9893-5b8ff3e52867-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.668639 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-df6468bbf-wcdr9"] Dec 04 06:16:51 crc kubenswrapper[4685]: I1204 06:16:51.675576 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-df6468bbf-wcdr9"] Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.679568 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f888497cf-2rw5z"] Dec 04 06:16:52 crc kubenswrapper[4685]: E1204 06:16:52.680238 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9774d34d-7e88-4dbd-9893-5b8ff3e52867" containerName="controller-manager" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.680261 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9774d34d-7e88-4dbd-9893-5b8ff3e52867" containerName="controller-manager" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.680452 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9774d34d-7e88-4dbd-9893-5b8ff3e52867" containerName="controller-manager" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.681020 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.686683 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.687681 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.687874 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.688186 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.688731 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.688804 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.695574 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.704203 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f888497cf-2rw5z"] Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.774089 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-client-ca\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.774148 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbc18019-1390-40de-9c92-850e90542066-serving-cert\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.774182 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5z96\" (UniqueName: \"kubernetes.io/projected/cbc18019-1390-40de-9c92-850e90542066-kube-api-access-k5z96\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.774212 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-config\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.774255 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-proxy-ca-bundles\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.875403 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5z96\" (UniqueName: \"kubernetes.io/projected/cbc18019-1390-40de-9c92-850e90542066-kube-api-access-k5z96\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.875846 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-config\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.876563 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-proxy-ca-bundles\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.876744 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-client-ca\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.876820 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbc18019-1390-40de-9c92-850e90542066-serving-cert\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.877381 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-client-ca\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.877727 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-proxy-ca-bundles\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.878256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc18019-1390-40de-9c92-850e90542066-config\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.885807 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbc18019-1390-40de-9c92-850e90542066-serving-cert\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.895509 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5z96\" (UniqueName: \"kubernetes.io/projected/cbc18019-1390-40de-9c92-850e90542066-kube-api-access-k5z96\") pod \"controller-manager-6f888497cf-2rw5z\" (UID: \"cbc18019-1390-40de-9c92-850e90542066\") " pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:52 crc kubenswrapper[4685]: I1204 06:16:52.996241 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:53 crc kubenswrapper[4685]: I1204 06:16:53.137034 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9774d34d-7e88-4dbd-9893-5b8ff3e52867" path="/var/lib/kubelet/pods/9774d34d-7e88-4dbd-9893-5b8ff3e52867/volumes" Dec 04 06:16:53 crc kubenswrapper[4685]: I1204 06:16:53.423727 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f888497cf-2rw5z"] Dec 04 06:16:54 crc kubenswrapper[4685]: I1204 06:16:54.351769 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" event={"ID":"cbc18019-1390-40de-9c92-850e90542066","Type":"ContainerStarted","Data":"9172b3b9d97d8fe835c69ab6b2e56cd0224eed096fc67d21d747afcaeb650847"} Dec 04 06:16:54 crc kubenswrapper[4685]: I1204 06:16:54.352186 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:54 crc kubenswrapper[4685]: I1204 06:16:54.352208 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" event={"ID":"cbc18019-1390-40de-9c92-850e90542066","Type":"ContainerStarted","Data":"712fa798a3090f978214c60469cbd4acb8451fb1f8657697201e317df4125b7b"} Dec 04 06:16:54 crc kubenswrapper[4685]: I1204 06:16:54.359971 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" Dec 04 06:16:54 crc kubenswrapper[4685]: I1204 06:16:54.375168 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f888497cf-2rw5z" podStartSLOduration=4.375148785 podStartE2EDuration="4.375148785s" podCreationTimestamp="2025-12-04 06:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:16:54.373926843 +0000 UTC m=+351.562157628" watchObservedRunningTime="2025-12-04 06:16:54.375148785 +0000 UTC m=+351.563379560" Dec 04 06:16:56 crc kubenswrapper[4685]: I1204 06:16:56.601486 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-2nf7t" Dec 04 06:16:56 crc kubenswrapper[4685]: I1204 06:16:56.670170 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jkj6h"] Dec 04 06:17:06 crc kubenswrapper[4685]: I1204 06:17:06.121305 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:17:06 crc kubenswrapper[4685]: I1204 06:17:06.122069 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:17:21 crc kubenswrapper[4685]: I1204 06:17:21.718529 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" podUID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" containerName="registry" containerID="cri-o://a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9" gracePeriod=30 Dec 04 06:17:21 crc kubenswrapper[4685]: I1204 06:17:21.742381 4685 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-jkj6h container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.16:5000/healthz\": dial tcp 10.217.0.16:5000: connect: connection refused" start-of-body= Dec 04 06:17:21 crc kubenswrapper[4685]: I1204 06:17:21.742949 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" podUID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.16:5000/healthz\": dial tcp 10.217.0.16:5000: connect: connection refused" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.157426 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275272 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-installation-pull-secrets\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275360 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-certificates\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275434 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-ca-trust-extracted\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275535 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-trusted-ca\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275862 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275899 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-tls\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.275958 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-bound-sa-token\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.276018 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6l48\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-kube-api-access-j6l48\") pod \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\" (UID: \"bfebae4b-a7c1-47a0-9efa-4c7127df9a13\") " Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.276610 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.280910 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.284566 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.284861 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.285109 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-kube-api-access-j6l48" (OuterVolumeSpecName: "kube-api-access-j6l48") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "kube-api-access-j6l48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.285726 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.290688 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.300802 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bfebae4b-a7c1-47a0-9efa-4c7127df9a13" (UID: "bfebae4b-a7c1-47a0-9efa-4c7127df9a13"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378372 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378469 4685 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378480 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378497 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6l48\" (UniqueName: \"kubernetes.io/projected/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-kube-api-access-j6l48\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378511 4685 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378519 4685 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.378528 4685 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfebae4b-a7c1-47a0-9efa-4c7127df9a13-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.544039 4685 generic.go:334] "Generic (PLEG): container finished" podID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" containerID="a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9" exitCode=0 Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.544130 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.544143 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" event={"ID":"bfebae4b-a7c1-47a0-9efa-4c7127df9a13","Type":"ContainerDied","Data":"a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9"} Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.544636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jkj6h" event={"ID":"bfebae4b-a7c1-47a0-9efa-4c7127df9a13","Type":"ContainerDied","Data":"2cca135bf3eda5692fa832aa05e27278bd2a5a5554ebc46110aeba0ca0c15cf6"} Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.544688 4685 scope.go:117] "RemoveContainer" containerID="a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.571038 4685 scope.go:117] "RemoveContainer" containerID="a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9" Dec 04 06:17:22 crc kubenswrapper[4685]: E1204 06:17:22.571609 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9\": container with ID starting with a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9 not found: ID does not exist" containerID="a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.571693 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9"} err="failed to get container status \"a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9\": rpc error: code = NotFound desc = could not find container \"a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9\": container with ID starting with a4f53d736042b54ea5e5ab9d148071880ff995826cb380781b1c89ec4ba238e9 not found: ID does not exist" Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.590272 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jkj6h"] Dec 04 06:17:22 crc kubenswrapper[4685]: I1204 06:17:22.593041 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jkj6h"] Dec 04 06:17:23 crc kubenswrapper[4685]: I1204 06:17:23.134613 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" path="/var/lib/kubelet/pods/bfebae4b-a7c1-47a0-9efa-4c7127df9a13/volumes" Dec 04 06:17:36 crc kubenswrapper[4685]: I1204 06:17:36.121312 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:17:36 crc kubenswrapper[4685]: I1204 06:17:36.122389 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.120931 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.121670 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.121730 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.122501 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c538fac20a9f6d91b3fbff5866d101a5f49b5694c6b6a972d11f5cc18b5b6a38"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.122596 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://c538fac20a9f6d91b3fbff5866d101a5f49b5694c6b6a972d11f5cc18b5b6a38" gracePeriod=600 Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.846992 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="c538fac20a9f6d91b3fbff5866d101a5f49b5694c6b6a972d11f5cc18b5b6a38" exitCode=0 Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.847090 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"c538fac20a9f6d91b3fbff5866d101a5f49b5694c6b6a972d11f5cc18b5b6a38"} Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.847868 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"2dda32d5375f2205695b926293374a6c3db5c6e9ff6115114489ad170e81d7dc"} Dec 04 06:18:06 crc kubenswrapper[4685]: I1204 06:18:06.847908 4685 scope.go:117] "RemoveContainer" containerID="97102e0f596e81ea6197a549615823edf36872d41188af0859134755005e8078" Dec 04 06:20:06 crc kubenswrapper[4685]: I1204 06:20:06.122041 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:20:06 crc kubenswrapper[4685]: I1204 06:20:06.122869 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:20:36 crc kubenswrapper[4685]: I1204 06:20:36.122079 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:20:36 crc kubenswrapper[4685]: I1204 06:20:36.123130 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:21:06 crc kubenswrapper[4685]: I1204 06:21:06.121816 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:21:06 crc kubenswrapper[4685]: I1204 06:21:06.123070 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:21:06 crc kubenswrapper[4685]: I1204 06:21:06.123154 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:21:06 crc kubenswrapper[4685]: I1204 06:21:06.124354 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2dda32d5375f2205695b926293374a6c3db5c6e9ff6115114489ad170e81d7dc"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:21:06 crc kubenswrapper[4685]: I1204 06:21:06.124463 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://2dda32d5375f2205695b926293374a6c3db5c6e9ff6115114489ad170e81d7dc" gracePeriod=600 Dec 04 06:21:07 crc kubenswrapper[4685]: I1204 06:21:07.225732 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="2dda32d5375f2205695b926293374a6c3db5c6e9ff6115114489ad170e81d7dc" exitCode=0 Dec 04 06:21:07 crc kubenswrapper[4685]: I1204 06:21:07.225835 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"2dda32d5375f2205695b926293374a6c3db5c6e9ff6115114489ad170e81d7dc"} Dec 04 06:21:07 crc kubenswrapper[4685]: I1204 06:21:07.226291 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"39d6b27b9ba56cb920b538ba51d04d90d1b790d65f04ecd9199b893b79b8e599"} Dec 04 06:21:07 crc kubenswrapper[4685]: I1204 06:21:07.226326 4685 scope.go:117] "RemoveContainer" containerID="c538fac20a9f6d91b3fbff5866d101a5f49b5694c6b6a972d11f5cc18b5b6a38" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.846916 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x4tmf"] Dec 04 06:22:28 crc kubenswrapper[4685]: E1204 06:22:28.847700 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" containerName="registry" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.847713 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" containerName="registry" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.847816 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfebae4b-a7c1-47a0-9efa-4c7127df9a13" containerName="registry" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.848154 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.851234 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.851263 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.851939 4685 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-pw4tj" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.860278 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nqsrd"] Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.861548 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nqsrd" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.865685 4685 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-d4qqp" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.873195 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x4tmf"] Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.879451 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-28r8x"] Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.880272 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.884199 4685 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-chrfd" Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.889560 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nqsrd"] Dec 04 06:22:28 crc kubenswrapper[4685]: I1204 06:22:28.897047 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-28r8x"] Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.036010 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqghk\" (UniqueName: \"kubernetes.io/projected/49a6e18b-01a4-44b9-8487-368f096cf059-kube-api-access-kqghk\") pod \"cert-manager-cainjector-7f985d654d-x4tmf\" (UID: \"49a6e18b-01a4-44b9-8487-368f096cf059\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.036082 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz9pd\" (UniqueName: \"kubernetes.io/projected/a33af606-6a04-4d1b-a5e5-b709163d4d89-kube-api-access-tz9pd\") pod \"cert-manager-webhook-5655c58dd6-28r8x\" (UID: \"a33af606-6a04-4d1b-a5e5-b709163d4d89\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.036171 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpqpg\" (UniqueName: \"kubernetes.io/projected/8ffffd54-b5c5-434a-b2d3-6afb6f12a254-kube-api-access-hpqpg\") pod \"cert-manager-5b446d88c5-nqsrd\" (UID: \"8ffffd54-b5c5-434a-b2d3-6afb6f12a254\") " pod="cert-manager/cert-manager-5b446d88c5-nqsrd" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.137119 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqghk\" (UniqueName: \"kubernetes.io/projected/49a6e18b-01a4-44b9-8487-368f096cf059-kube-api-access-kqghk\") pod \"cert-manager-cainjector-7f985d654d-x4tmf\" (UID: \"49a6e18b-01a4-44b9-8487-368f096cf059\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.137186 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz9pd\" (UniqueName: \"kubernetes.io/projected/a33af606-6a04-4d1b-a5e5-b709163d4d89-kube-api-access-tz9pd\") pod \"cert-manager-webhook-5655c58dd6-28r8x\" (UID: \"a33af606-6a04-4d1b-a5e5-b709163d4d89\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.137249 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpqpg\" (UniqueName: \"kubernetes.io/projected/8ffffd54-b5c5-434a-b2d3-6afb6f12a254-kube-api-access-hpqpg\") pod \"cert-manager-5b446d88c5-nqsrd\" (UID: \"8ffffd54-b5c5-434a-b2d3-6afb6f12a254\") " pod="cert-manager/cert-manager-5b446d88c5-nqsrd" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.155881 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqghk\" (UniqueName: \"kubernetes.io/projected/49a6e18b-01a4-44b9-8487-368f096cf059-kube-api-access-kqghk\") pod \"cert-manager-cainjector-7f985d654d-x4tmf\" (UID: \"49a6e18b-01a4-44b9-8487-368f096cf059\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.156218 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz9pd\" (UniqueName: \"kubernetes.io/projected/a33af606-6a04-4d1b-a5e5-b709163d4d89-kube-api-access-tz9pd\") pod \"cert-manager-webhook-5655c58dd6-28r8x\" (UID: \"a33af606-6a04-4d1b-a5e5-b709163d4d89\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.159703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpqpg\" (UniqueName: \"kubernetes.io/projected/8ffffd54-b5c5-434a-b2d3-6afb6f12a254-kube-api-access-hpqpg\") pod \"cert-manager-5b446d88c5-nqsrd\" (UID: \"8ffffd54-b5c5-434a-b2d3-6afb6f12a254\") " pod="cert-manager/cert-manager-5b446d88c5-nqsrd" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.163294 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.176362 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nqsrd" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.194040 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.406147 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x4tmf"] Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.415784 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.455162 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nqsrd"] Dec 04 06:22:29 crc kubenswrapper[4685]: W1204 06:22:29.462778 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ffffd54_b5c5_434a_b2d3_6afb6f12a254.slice/crio-01d2ce0dbcad26123404e6a0ed36796373fdf8f54c855092641df584af5de02b WatchSource:0}: Error finding container 01d2ce0dbcad26123404e6a0ed36796373fdf8f54c855092641df584af5de02b: Status 404 returned error can't find the container with id 01d2ce0dbcad26123404e6a0ed36796373fdf8f54c855092641df584af5de02b Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.488462 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-28r8x"] Dec 04 06:22:29 crc kubenswrapper[4685]: W1204 06:22:29.500583 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda33af606_6a04_4d1b_a5e5_b709163d4d89.slice/crio-14bfac08c1d136798ac788c989be2303f75367a36bc52344057bc0060cda4774 WatchSource:0}: Error finding container 14bfac08c1d136798ac788c989be2303f75367a36bc52344057bc0060cda4774: Status 404 returned error can't find the container with id 14bfac08c1d136798ac788c989be2303f75367a36bc52344057bc0060cda4774 Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.811684 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nqsrd" event={"ID":"8ffffd54-b5c5-434a-b2d3-6afb6f12a254","Type":"ContainerStarted","Data":"01d2ce0dbcad26123404e6a0ed36796373fdf8f54c855092641df584af5de02b"} Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.813169 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" event={"ID":"49a6e18b-01a4-44b9-8487-368f096cf059","Type":"ContainerStarted","Data":"0f2f36d844bb7b2d96912082cfe50fb487b2e157514b1e947e9ea17c4a34bb2b"} Dec 04 06:22:29 crc kubenswrapper[4685]: I1204 06:22:29.814651 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" event={"ID":"a33af606-6a04-4d1b-a5e5-b709163d4d89","Type":"ContainerStarted","Data":"14bfac08c1d136798ac788c989be2303f75367a36bc52344057bc0060cda4774"} Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.847795 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" event={"ID":"49a6e18b-01a4-44b9-8487-368f096cf059","Type":"ContainerStarted","Data":"a14b9ae71d8f89ba173b2fa8942c6b74951799203dc5300eb2c93ab0cbb67ad0"} Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.858139 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" event={"ID":"a33af606-6a04-4d1b-a5e5-b709163d4d89","Type":"ContainerStarted","Data":"2103c75b420ac3bdc2e769245c6c961fefd3b33a9d74844e42145f402a2b3891"} Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.858548 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.864084 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nqsrd" event={"ID":"8ffffd54-b5c5-434a-b2d3-6afb6f12a254","Type":"ContainerStarted","Data":"dbeffdcb59682bdd7d9652000e1245f2a3d187e0141d009a6c256b7e3a4fb252"} Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.880486 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-x4tmf" podStartSLOduration=2.372577067 podStartE2EDuration="5.880377832s" podCreationTimestamp="2025-12-04 06:22:28 +0000 UTC" firstStartedPulling="2025-12-04 06:22:29.415512348 +0000 UTC m=+686.603743123" lastFinishedPulling="2025-12-04 06:22:32.923313073 +0000 UTC m=+690.111543888" observedRunningTime="2025-12-04 06:22:33.870865237 +0000 UTC m=+691.059096082" watchObservedRunningTime="2025-12-04 06:22:33.880377832 +0000 UTC m=+691.068608597" Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.900892 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nqsrd" podStartSLOduration=2.449039064 podStartE2EDuration="5.900862398s" podCreationTimestamp="2025-12-04 06:22:28 +0000 UTC" firstStartedPulling="2025-12-04 06:22:29.46527558 +0000 UTC m=+686.653526186" lastFinishedPulling="2025-12-04 06:22:32.917118705 +0000 UTC m=+690.105349520" observedRunningTime="2025-12-04 06:22:33.889807704 +0000 UTC m=+691.078038499" watchObservedRunningTime="2025-12-04 06:22:33.900862398 +0000 UTC m=+691.089093193" Dec 04 06:22:33 crc kubenswrapper[4685]: I1204 06:22:33.913509 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" podStartSLOduration=2.493254287 podStartE2EDuration="5.913479321s" podCreationTimestamp="2025-12-04 06:22:28 +0000 UTC" firstStartedPulling="2025-12-04 06:22:29.503270535 +0000 UTC m=+686.691501310" lastFinishedPulling="2025-12-04 06:22:32.923495559 +0000 UTC m=+690.111726344" observedRunningTime="2025-12-04 06:22:33.911093885 +0000 UTC m=+691.099324690" watchObservedRunningTime="2025-12-04 06:22:33.913479321 +0000 UTC m=+691.101710126" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.197554 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-28r8x" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.489264 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n4rjr"] Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.489982 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-controller" containerID="cri-o://f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.490020 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="sbdb" containerID="cri-o://6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.490149 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-acl-logging" containerID="cri-o://9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.490140 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-node" containerID="cri-o://01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.490140 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.490196 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="nbdb" containerID="cri-o://d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.490231 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="northd" containerID="cri-o://9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.568725 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" containerID="cri-o://d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" gracePeriod=30 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.874788 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/3.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.878010 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovn-acl-logging/0.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.878662 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovn-controller/0.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.879249 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.904954 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/2.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.905925 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/1.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.905988 4685 generic.go:334] "Generic (PLEG): container finished" podID="736d38b0-c0ce-4db5-bee6-72ee0dc23b40" containerID="23cbb5fd93c677593365a3111be3cf45e952008b1d7520ce81f4f2b7c902f1eb" exitCode=2 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.906022 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerDied","Data":"23cbb5fd93c677593365a3111be3cf45e952008b1d7520ce81f4f2b7c902f1eb"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.906071 4685 scope.go:117] "RemoveContainer" containerID="b77c3271ceedbfd70f85cb3869c47f87941fc52ce44dd930c38bd4119d807643" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.906726 4685 scope.go:117] "RemoveContainer" containerID="23cbb5fd93c677593365a3111be3cf45e952008b1d7520ce81f4f2b7c902f1eb" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.907140 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dpd9d_openshift-multus(736d38b0-c0ce-4db5-bee6-72ee0dc23b40)\"" pod="openshift-multus/multus-dpd9d" podUID="736d38b0-c0ce-4db5-bee6-72ee0dc23b40" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.910967 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovnkube-controller/3.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.918006 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovn-acl-logging/0.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919249 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n4rjr_f4c7de36-abbd-448e-bf66-eecafc38068c/ovn-controller/0.log" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919828 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" exitCode=0 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919869 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" exitCode=0 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919882 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" exitCode=0 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919895 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" exitCode=0 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919904 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" exitCode=0 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919915 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" exitCode=0 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919927 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" exitCode=143 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919937 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" exitCode=143 Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919940 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.919964 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920002 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920020 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920035 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920049 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920064 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920079 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920093 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920100 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920107 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920114 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920121 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920132 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920139 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920146 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920153 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920165 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920175 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920183 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920190 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920197 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920214 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920221 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920229 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920236 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920242 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920251 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920264 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920278 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920288 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920297 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920305 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920313 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920320 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920327 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920335 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920343 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920350 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920359 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n4rjr" event={"ID":"f4c7de36-abbd-448e-bf66-eecafc38068c","Type":"ContainerDied","Data":"69c706cf67d0a5846a9f866fec57c909e9894067daa991ebecda10951dc2192e"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920370 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920379 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920386 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920393 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920401 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920429 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920438 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920445 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920452 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.920459 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946239 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cvv84"] Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946461 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kubecfg-setup" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946473 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kubecfg-setup" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946485 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946490 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946501 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946506 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946514 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-node" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946520 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-node" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946527 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="nbdb" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946534 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="nbdb" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946544 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="northd" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946551 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="northd" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946558 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946564 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946570 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="sbdb" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946576 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="sbdb" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946588 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946593 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946601 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946607 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946617 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-acl-logging" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946623 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-acl-logging" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946628 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946635 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946743 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946756 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="kube-rbac-proxy-node" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946766 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946774 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="nbdb" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946783 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-acl-logging" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946791 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovn-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946800 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946806 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="northd" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946812 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946819 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946829 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="sbdb" Dec 04 06:22:39 crc kubenswrapper[4685]: E1204 06:22:39.946925 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.946936 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.947051 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" containerName="ovnkube-controller" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.948939 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.954316 4685 scope.go:117] "RemoveContainer" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" Dec 04 06:22:39 crc kubenswrapper[4685]: I1204 06:22:39.978913 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.004724 4685 scope.go:117] "RemoveContainer" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-ovn-kubernetes\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017598 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-config\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017626 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-log-socket\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017634 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4c7de36-abbd-448e-bf66-eecafc38068c-ovn-node-metrics-cert\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017693 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-netns\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017712 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-openvswitch\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017711 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-log-socket" (OuterVolumeSpecName: "log-socket") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017732 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-systemd-units\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017759 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l97ms\" (UniqueName: \"kubernetes.io/projected/f4c7de36-abbd-448e-bf66-eecafc38068c-kube-api-access-l97ms\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017777 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-etc-openvswitch\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017809 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-env-overrides\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017836 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-ovn\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017866 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017891 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-slash\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017918 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-script-lib\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017939 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-systemd\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017968 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-bin\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.017990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-node-log\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018025 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-kubelet\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018044 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-var-lib-openvswitch\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018065 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-netd\") pod \"f4c7de36-abbd-448e-bf66-eecafc38068c\" (UID: \"f4c7de36-abbd-448e-bf66-eecafc38068c\") " Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018222 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018287 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018421 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018473 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018501 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018523 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018565 4685 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018590 4685 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018605 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018618 4685 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018658 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018711 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018739 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-slash" (OuterVolumeSpecName: "host-slash") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.018929 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-node-log" (OuterVolumeSpecName: "node-log") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.019034 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.019043 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.019061 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.019204 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.019636 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.025104 4685 scope.go:117] "RemoveContainer" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.025538 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4c7de36-abbd-448e-bf66-eecafc38068c-kube-api-access-l97ms" (OuterVolumeSpecName: "kube-api-access-l97ms") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "kube-api-access-l97ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.026301 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c7de36-abbd-448e-bf66-eecafc38068c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.036268 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f4c7de36-abbd-448e-bf66-eecafc38068c" (UID: "f4c7de36-abbd-448e-bf66-eecafc38068c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.041740 4685 scope.go:117] "RemoveContainer" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.062693 4685 scope.go:117] "RemoveContainer" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.080519 4685 scope.go:117] "RemoveContainer" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.100145 4685 scope.go:117] "RemoveContainer" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120075 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-systemd-units\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120144 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-ovnkube-config\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120197 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-ovn\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120660 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-var-lib-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120712 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-ovnkube-script-lib\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120752 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-slash\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120782 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-node-log\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120805 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-log-socket\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120828 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.120986 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9k8h\" (UniqueName: \"kubernetes.io/projected/180160a0-752f-4217-8b38-995f55eeff01-kube-api-access-v9k8h\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121057 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-etc-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121095 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/180160a0-752f-4217-8b38-995f55eeff01-ovn-node-metrics-cert\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121176 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121222 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-run-netns\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121268 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-env-overrides\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121313 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-systemd\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121365 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-cni-netd\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121630 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-cni-bin\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121741 4685 scope.go:117] "RemoveContainer" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121826 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-kubelet\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121945 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4c7de36-abbd-448e-bf66-eecafc38068c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121977 4685 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.121996 4685 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122013 4685 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122033 4685 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122053 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l97ms\" (UniqueName: \"kubernetes.io/projected/f4c7de36-abbd-448e-bf66-eecafc38068c-kube-api-access-l97ms\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122071 4685 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122088 4685 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122112 4685 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122130 4685 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122150 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f4c7de36-abbd-448e-bf66-eecafc38068c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122168 4685 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122186 4685 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122202 4685 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122261 4685 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.122279 4685 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f4c7de36-abbd-448e-bf66-eecafc38068c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.146367 4685 scope.go:117] "RemoveContainer" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.167826 4685 scope.go:117] "RemoveContainer" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.168650 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": container with ID starting with d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac not found: ID does not exist" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.168721 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} err="failed to get container status \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": rpc error: code = NotFound desc = could not find container \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": container with ID starting with d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.168774 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.169272 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": container with ID starting with c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4 not found: ID does not exist" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.169316 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} err="failed to get container status \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": rpc error: code = NotFound desc = could not find container \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": container with ID starting with c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.169347 4685 scope.go:117] "RemoveContainer" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.169765 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": container with ID starting with 6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2 not found: ID does not exist" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.169839 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} err="failed to get container status \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": rpc error: code = NotFound desc = could not find container \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": container with ID starting with 6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.169873 4685 scope.go:117] "RemoveContainer" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.170335 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": container with ID starting with d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556 not found: ID does not exist" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.170374 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} err="failed to get container status \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": rpc error: code = NotFound desc = could not find container \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": container with ID starting with d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.170399 4685 scope.go:117] "RemoveContainer" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.171099 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": container with ID starting with 9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a not found: ID does not exist" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.171138 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} err="failed to get container status \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": rpc error: code = NotFound desc = could not find container \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": container with ID starting with 9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.171163 4685 scope.go:117] "RemoveContainer" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.172046 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": container with ID starting with 8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b not found: ID does not exist" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.172109 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} err="failed to get container status \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": rpc error: code = NotFound desc = could not find container \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": container with ID starting with 8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.172152 4685 scope.go:117] "RemoveContainer" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.172628 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": container with ID starting with 01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970 not found: ID does not exist" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.172667 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} err="failed to get container status \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": rpc error: code = NotFound desc = could not find container \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": container with ID starting with 01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.172693 4685 scope.go:117] "RemoveContainer" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.173086 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": container with ID starting with 9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c not found: ID does not exist" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.173146 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} err="failed to get container status \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": rpc error: code = NotFound desc = could not find container \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": container with ID starting with 9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.173183 4685 scope.go:117] "RemoveContainer" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.173631 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": container with ID starting with f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f not found: ID does not exist" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.173683 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} err="failed to get container status \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": rpc error: code = NotFound desc = could not find container \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": container with ID starting with f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.173717 4685 scope.go:117] "RemoveContainer" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" Dec 04 06:22:40 crc kubenswrapper[4685]: E1204 06:22:40.174100 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": container with ID starting with a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911 not found: ID does not exist" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.174134 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} err="failed to get container status \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": rpc error: code = NotFound desc = could not find container \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": container with ID starting with a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.174160 4685 scope.go:117] "RemoveContainer" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.174549 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} err="failed to get container status \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": rpc error: code = NotFound desc = could not find container \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": container with ID starting with d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.174602 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.175162 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} err="failed to get container status \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": rpc error: code = NotFound desc = could not find container \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": container with ID starting with c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.175216 4685 scope.go:117] "RemoveContainer" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.175617 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} err="failed to get container status \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": rpc error: code = NotFound desc = could not find container \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": container with ID starting with 6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.175666 4685 scope.go:117] "RemoveContainer" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.175974 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} err="failed to get container status \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": rpc error: code = NotFound desc = could not find container \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": container with ID starting with d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.176011 4685 scope.go:117] "RemoveContainer" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.176303 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} err="failed to get container status \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": rpc error: code = NotFound desc = could not find container \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": container with ID starting with 9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.176341 4685 scope.go:117] "RemoveContainer" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.176660 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} err="failed to get container status \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": rpc error: code = NotFound desc = could not find container \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": container with ID starting with 8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.176688 4685 scope.go:117] "RemoveContainer" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.176994 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} err="failed to get container status \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": rpc error: code = NotFound desc = could not find container \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": container with ID starting with 01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.177020 4685 scope.go:117] "RemoveContainer" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.177358 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} err="failed to get container status \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": rpc error: code = NotFound desc = could not find container \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": container with ID starting with 9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.177383 4685 scope.go:117] "RemoveContainer" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.177753 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} err="failed to get container status \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": rpc error: code = NotFound desc = could not find container \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": container with ID starting with f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.177781 4685 scope.go:117] "RemoveContainer" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.178044 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} err="failed to get container status \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": rpc error: code = NotFound desc = could not find container \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": container with ID starting with a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.178067 4685 scope.go:117] "RemoveContainer" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.178384 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} err="failed to get container status \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": rpc error: code = NotFound desc = could not find container \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": container with ID starting with d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.178435 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.178914 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} err="failed to get container status \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": rpc error: code = NotFound desc = could not find container \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": container with ID starting with c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.178982 4685 scope.go:117] "RemoveContainer" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.179582 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} err="failed to get container status \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": rpc error: code = NotFound desc = could not find container \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": container with ID starting with 6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.179641 4685 scope.go:117] "RemoveContainer" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.180155 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} err="failed to get container status \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": rpc error: code = NotFound desc = could not find container \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": container with ID starting with d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.180183 4685 scope.go:117] "RemoveContainer" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.180723 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} err="failed to get container status \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": rpc error: code = NotFound desc = could not find container \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": container with ID starting with 9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.180777 4685 scope.go:117] "RemoveContainer" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.181168 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} err="failed to get container status \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": rpc error: code = NotFound desc = could not find container \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": container with ID starting with 8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.181197 4685 scope.go:117] "RemoveContainer" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.181507 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} err="failed to get container status \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": rpc error: code = NotFound desc = could not find container \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": container with ID starting with 01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.181540 4685 scope.go:117] "RemoveContainer" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.181837 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} err="failed to get container status \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": rpc error: code = NotFound desc = could not find container \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": container with ID starting with 9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.181871 4685 scope.go:117] "RemoveContainer" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.182191 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} err="failed to get container status \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": rpc error: code = NotFound desc = could not find container \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": container with ID starting with f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.182228 4685 scope.go:117] "RemoveContainer" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.182614 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} err="failed to get container status \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": rpc error: code = NotFound desc = could not find container \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": container with ID starting with a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.182652 4685 scope.go:117] "RemoveContainer" containerID="d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.183014 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac"} err="failed to get container status \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": rpc error: code = NotFound desc = could not find container \"d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac\": container with ID starting with d6ce0498cc94c78abde0a865710783f877c0b6d78305bf2d808a4590daec45ac not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.183050 4685 scope.go:117] "RemoveContainer" containerID="c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.183444 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4"} err="failed to get container status \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": rpc error: code = NotFound desc = could not find container \"c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4\": container with ID starting with c30f92b647825434a328550273df9a9ae2843450f8e03206a86f453d5f62cce4 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.183477 4685 scope.go:117] "RemoveContainer" containerID="6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.183763 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2"} err="failed to get container status \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": rpc error: code = NotFound desc = could not find container \"6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2\": container with ID starting with 6fcb1a66aed42ec3e8ca77cdc58dfe87ad229459a5f53e3de8e4f22ff13618a2 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.183797 4685 scope.go:117] "RemoveContainer" containerID="d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.184118 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556"} err="failed to get container status \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": rpc error: code = NotFound desc = could not find container \"d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556\": container with ID starting with d00a702561a341b97c16ed0c616530c601eab480f149832c8c05b96806de9556 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.184155 4685 scope.go:117] "RemoveContainer" containerID="9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.184477 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a"} err="failed to get container status \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": rpc error: code = NotFound desc = could not find container \"9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a\": container with ID starting with 9aa2c4403cff884bf631883cbe961f09edcebdc364ee3a1c0fe2d9aeac97d51a not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.184505 4685 scope.go:117] "RemoveContainer" containerID="8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.184846 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b"} err="failed to get container status \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": rpc error: code = NotFound desc = could not find container \"8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b\": container with ID starting with 8467505068be40b278815b0e943f6edd6ada1ad0ba17e8b9d7ec7340cda2f58b not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.184888 4685 scope.go:117] "RemoveContainer" containerID="01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.185232 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970"} err="failed to get container status \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": rpc error: code = NotFound desc = could not find container \"01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970\": container with ID starting with 01c7ad20819c4b314716fbf577221df8b3a9e315bee89913240bc9587c896970 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.185291 4685 scope.go:117] "RemoveContainer" containerID="9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.185679 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c"} err="failed to get container status \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": rpc error: code = NotFound desc = could not find container \"9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c\": container with ID starting with 9835c7a2dc22eb68b1b5d824ae028e06ca5a13b71676a4f723d982920b1e9b1c not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.185717 4685 scope.go:117] "RemoveContainer" containerID="f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.186047 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f"} err="failed to get container status \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": rpc error: code = NotFound desc = could not find container \"f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f\": container with ID starting with f58648328a265bde0a586d36547c0d11a0de97ccd7620a9697f6e1dea56a274f not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.186084 4685 scope.go:117] "RemoveContainer" containerID="a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.186582 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911"} err="failed to get container status \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": rpc error: code = NotFound desc = could not find container \"a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911\": container with ID starting with a90ba7c7c5a29af09558ee8892d1dc5cc046dc4498e308da2e485e5b63681911 not found: ID does not exist" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.222828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.222892 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.222925 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9k8h\" (UniqueName: \"kubernetes.io/projected/180160a0-752f-4217-8b38-995f55eeff01-kube-api-access-v9k8h\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.222963 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.222985 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-etc-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223046 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-etc-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223078 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/180160a0-752f-4217-8b38-995f55eeff01-ovn-node-metrics-cert\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223117 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223246 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223303 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-run-netns\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223366 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-env-overrides\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223441 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-systemd\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223484 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-cni-netd\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223545 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-cni-bin\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223557 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-run-netns\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223613 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-kubelet\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223751 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223788 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-cni-netd\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223792 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-systemd-units\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223796 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-systemd\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223808 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-cni-bin\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223847 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-kubelet\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.223877 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-systemd-units\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-ovnkube-config\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224054 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-ovn\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224074 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-var-lib-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224094 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-ovnkube-script-lib\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224119 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-slash\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224140 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-node-log\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224136 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-run-ovn\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224159 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-log-socket\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224248 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-log-socket\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224280 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-host-slash\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224307 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-node-log\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224334 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/180160a0-752f-4217-8b38-995f55eeff01-var-lib-openvswitch\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.224945 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-env-overrides\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.225038 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-ovnkube-config\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.225621 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/180160a0-752f-4217-8b38-995f55eeff01-ovnkube-script-lib\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.230077 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/180160a0-752f-4217-8b38-995f55eeff01-ovn-node-metrics-cert\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.256980 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9k8h\" (UniqueName: \"kubernetes.io/projected/180160a0-752f-4217-8b38-995f55eeff01-kube-api-access-v9k8h\") pod \"ovnkube-node-cvv84\" (UID: \"180160a0-752f-4217-8b38-995f55eeff01\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.274652 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.280942 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n4rjr"] Dec 04 06:22:40 crc kubenswrapper[4685]: I1204 06:22:40.284683 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n4rjr"] Dec 04 06:22:40 crc kubenswrapper[4685]: W1204 06:22:40.299852 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod180160a0_752f_4217_8b38_995f55eeff01.slice/crio-b6044c49a383e73f1ba23951a4b1c029fc883f586742cf1ee9c58f367ea235d4 WatchSource:0}: Error finding container b6044c49a383e73f1ba23951a4b1c029fc883f586742cf1ee9c58f367ea235d4: Status 404 returned error can't find the container with id b6044c49a383e73f1ba23951a4b1c029fc883f586742cf1ee9c58f367ea235d4 Dec 04 06:22:41 crc kubenswrapper[4685]: I1204 06:22:41.749943 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4c7de36-abbd-448e-bf66-eecafc38068c" path="/var/lib/kubelet/pods/f4c7de36-abbd-448e-bf66-eecafc38068c/volumes" Dec 04 06:22:41 crc kubenswrapper[4685]: I1204 06:22:41.752299 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/2.log" Dec 04 06:22:41 crc kubenswrapper[4685]: I1204 06:22:41.759966 4685 generic.go:334] "Generic (PLEG): container finished" podID="180160a0-752f-4217-8b38-995f55eeff01" containerID="6e6209c9a3b7ebd4b36ac1dda037b05fa2b725d2a1f3e82d67f91e297b6189b9" exitCode=0 Dec 04 06:22:41 crc kubenswrapper[4685]: I1204 06:22:41.760029 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerDied","Data":"6e6209c9a3b7ebd4b36ac1dda037b05fa2b725d2a1f3e82d67f91e297b6189b9"} Dec 04 06:22:41 crc kubenswrapper[4685]: I1204 06:22:41.760068 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"b6044c49a383e73f1ba23951a4b1c029fc883f586742cf1ee9c58f367ea235d4"} Dec 04 06:22:42 crc kubenswrapper[4685]: I1204 06:22:42.774350 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"af31bc788881298d420da1835b929a4d9e7a26e2664ee4d0fe1a0b73de9ac2e1"} Dec 04 06:22:42 crc kubenswrapper[4685]: I1204 06:22:42.775247 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"c7bbc049c43de5f989151b5857d83d3c3df6ca422b41a7218f9daedcd3a20ebe"} Dec 04 06:22:42 crc kubenswrapper[4685]: I1204 06:22:42.775265 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"7ed74ce26340bcbbf5c0a58aac60c7e50c970d24a259861cd55adb4925cd9e95"} Dec 04 06:22:42 crc kubenswrapper[4685]: I1204 06:22:42.775294 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"e16cacef45d7a40c61e9c07118ea2805836c33981d1c5cc4dc965bbd4157dbeb"} Dec 04 06:22:42 crc kubenswrapper[4685]: I1204 06:22:42.775309 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"5468880c2fe2bb27e6be7093fefc8b7ae1448a45981fd25fba22be0c4d4dab3b"} Dec 04 06:22:42 crc kubenswrapper[4685]: I1204 06:22:42.775320 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"b2c2608af93193ff6b0788d58ee0d1d798141094a7dddc004e85156efbc3bf6f"} Dec 04 06:22:45 crc kubenswrapper[4685]: I1204 06:22:45.798634 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"8935a2f92a9ad844b7ab1200699a4705ea4e27c1a9d8edb2289221463424cf06"} Dec 04 06:22:47 crc kubenswrapper[4685]: I1204 06:22:47.817507 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" event={"ID":"180160a0-752f-4217-8b38-995f55eeff01","Type":"ContainerStarted","Data":"e8a4008d816d247375a4f3b526ae6cc0556fdb16aa7446bd10eab7573abcedd5"} Dec 04 06:22:47 crc kubenswrapper[4685]: I1204 06:22:47.818005 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:47 crc kubenswrapper[4685]: I1204 06:22:47.818022 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:47 crc kubenswrapper[4685]: I1204 06:22:47.854154 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" podStartSLOduration=8.854136932 podStartE2EDuration="8.854136932s" podCreationTimestamp="2025-12-04 06:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:22:47.852631714 +0000 UTC m=+705.040862489" watchObservedRunningTime="2025-12-04 06:22:47.854136932 +0000 UTC m=+705.042367707" Dec 04 06:22:47 crc kubenswrapper[4685]: I1204 06:22:47.861799 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:48 crc kubenswrapper[4685]: I1204 06:22:48.825775 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:48 crc kubenswrapper[4685]: I1204 06:22:48.892838 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:22:52 crc kubenswrapper[4685]: I1204 06:22:52.125959 4685 scope.go:117] "RemoveContainer" containerID="23cbb5fd93c677593365a3111be3cf45e952008b1d7520ce81f4f2b7c902f1eb" Dec 04 06:22:52 crc kubenswrapper[4685]: E1204 06:22:52.126551 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dpd9d_openshift-multus(736d38b0-c0ce-4db5-bee6-72ee0dc23b40)\"" pod="openshift-multus/multus-dpd9d" podUID="736d38b0-c0ce-4db5-bee6-72ee0dc23b40" Dec 04 06:23:05 crc kubenswrapper[4685]: I1204 06:23:05.125898 4685 scope.go:117] "RemoveContainer" containerID="23cbb5fd93c677593365a3111be3cf45e952008b1d7520ce81f4f2b7c902f1eb" Dec 04 06:23:05 crc kubenswrapper[4685]: I1204 06:23:05.949570 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dpd9d_736d38b0-c0ce-4db5-bee6-72ee0dc23b40/kube-multus/2.log" Dec 04 06:23:05 crc kubenswrapper[4685]: I1204 06:23:05.950092 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dpd9d" event={"ID":"736d38b0-c0ce-4db5-bee6-72ee0dc23b40","Type":"ContainerStarted","Data":"abbaa66d36057390e85b08c033e340eff7ef5468ebde257f1e9518b46a752a20"} Dec 04 06:23:06 crc kubenswrapper[4685]: I1204 06:23:06.120867 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:23:06 crc kubenswrapper[4685]: I1204 06:23:06.120963 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:23:10 crc kubenswrapper[4685]: I1204 06:23:10.316889 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvv84" Dec 04 06:23:23 crc kubenswrapper[4685]: I1204 06:23:23.941491 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp"] Dec 04 06:23:23 crc kubenswrapper[4685]: I1204 06:23:23.945429 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:23 crc kubenswrapper[4685]: I1204 06:23:23.953262 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 06:23:23 crc kubenswrapper[4685]: I1204 06:23:23.969663 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp"] Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.003820 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.003926 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.003982 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfwng\" (UniqueName: \"kubernetes.io/projected/05790b45-60d8-406e-857d-0f85d286cdb9-kube-api-access-gfwng\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.105313 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfwng\" (UniqueName: \"kubernetes.io/projected/05790b45-60d8-406e-857d-0f85d286cdb9-kube-api-access-gfwng\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.105434 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.105473 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.106009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.106142 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.126398 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfwng\" (UniqueName: \"kubernetes.io/projected/05790b45-60d8-406e-857d-0f85d286cdb9-kube-api-access-gfwng\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.270037 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:24 crc kubenswrapper[4685]: I1204 06:23:24.568567 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp"] Dec 04 06:23:25 crc kubenswrapper[4685]: I1204 06:23:25.095752 4685 generic.go:334] "Generic (PLEG): container finished" podID="05790b45-60d8-406e-857d-0f85d286cdb9" containerID="cf2bd52b9e9d47906a982ab97a0faa68706269fe219b27744de5443f806aba5a" exitCode=0 Dec 04 06:23:25 crc kubenswrapper[4685]: I1204 06:23:25.095822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" event={"ID":"05790b45-60d8-406e-857d-0f85d286cdb9","Type":"ContainerDied","Data":"cf2bd52b9e9d47906a982ab97a0faa68706269fe219b27744de5443f806aba5a"} Dec 04 06:23:25 crc kubenswrapper[4685]: I1204 06:23:25.095861 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" event={"ID":"05790b45-60d8-406e-857d-0f85d286cdb9","Type":"ContainerStarted","Data":"1e594956c43420b0e10da7ebf14aaf3bef470c834576d5a1da6d559d62aebc7f"} Dec 04 06:23:27 crc kubenswrapper[4685]: I1204 06:23:27.108876 4685 generic.go:334] "Generic (PLEG): container finished" podID="05790b45-60d8-406e-857d-0f85d286cdb9" containerID="26bd6b6991ca6b226b2e207fb037f99907891c96a1b68976f90e4d7d3d70c94b" exitCode=0 Dec 04 06:23:27 crc kubenswrapper[4685]: I1204 06:23:27.108936 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" event={"ID":"05790b45-60d8-406e-857d-0f85d286cdb9","Type":"ContainerDied","Data":"26bd6b6991ca6b226b2e207fb037f99907891c96a1b68976f90e4d7d3d70c94b"} Dec 04 06:23:28 crc kubenswrapper[4685]: I1204 06:23:28.130211 4685 generic.go:334] "Generic (PLEG): container finished" podID="05790b45-60d8-406e-857d-0f85d286cdb9" containerID="67149061ff70f731b5c112f4d72a52dfff424dd09a14a0328728201282b1e0a7" exitCode=0 Dec 04 06:23:28 crc kubenswrapper[4685]: I1204 06:23:28.130277 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" event={"ID":"05790b45-60d8-406e-857d-0f85d286cdb9","Type":"ContainerDied","Data":"67149061ff70f731b5c112f4d72a52dfff424dd09a14a0328728201282b1e0a7"} Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.463439 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.488772 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfwng\" (UniqueName: \"kubernetes.io/projected/05790b45-60d8-406e-857d-0f85d286cdb9-kube-api-access-gfwng\") pod \"05790b45-60d8-406e-857d-0f85d286cdb9\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.488882 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-bundle\") pod \"05790b45-60d8-406e-857d-0f85d286cdb9\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.488933 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-util\") pod \"05790b45-60d8-406e-857d-0f85d286cdb9\" (UID: \"05790b45-60d8-406e-857d-0f85d286cdb9\") " Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.490006 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-bundle" (OuterVolumeSpecName: "bundle") pod "05790b45-60d8-406e-857d-0f85d286cdb9" (UID: "05790b45-60d8-406e-857d-0f85d286cdb9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.498019 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05790b45-60d8-406e-857d-0f85d286cdb9-kube-api-access-gfwng" (OuterVolumeSpecName: "kube-api-access-gfwng") pod "05790b45-60d8-406e-857d-0f85d286cdb9" (UID: "05790b45-60d8-406e-857d-0f85d286cdb9"). InnerVolumeSpecName "kube-api-access-gfwng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.527309 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-util" (OuterVolumeSpecName: "util") pod "05790b45-60d8-406e-857d-0f85d286cdb9" (UID: "05790b45-60d8-406e-857d-0f85d286cdb9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.590104 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfwng\" (UniqueName: \"kubernetes.io/projected/05790b45-60d8-406e-857d-0f85d286cdb9-kube-api-access-gfwng\") on node \"crc\" DevicePath \"\"" Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.590188 4685 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:23:29 crc kubenswrapper[4685]: I1204 06:23:29.590209 4685 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/05790b45-60d8-406e-857d-0f85d286cdb9-util\") on node \"crc\" DevicePath \"\"" Dec 04 06:23:30 crc kubenswrapper[4685]: I1204 06:23:30.151986 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" event={"ID":"05790b45-60d8-406e-857d-0f85d286cdb9","Type":"ContainerDied","Data":"1e594956c43420b0e10da7ebf14aaf3bef470c834576d5a1da6d559d62aebc7f"} Dec 04 06:23:30 crc kubenswrapper[4685]: I1204 06:23:30.152348 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e594956c43420b0e10da7ebf14aaf3bef470c834576d5a1da6d559d62aebc7f" Dec 04 06:23:30 crc kubenswrapper[4685]: I1204 06:23:30.152150 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.501943 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488"] Dec 04 06:23:32 crc kubenswrapper[4685]: E1204 06:23:32.502138 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="util" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.502150 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="util" Dec 04 06:23:32 crc kubenswrapper[4685]: E1204 06:23:32.502162 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="extract" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.502169 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="extract" Dec 04 06:23:32 crc kubenswrapper[4685]: E1204 06:23:32.502182 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="pull" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.502189 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="pull" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.502292 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="05790b45-60d8-406e-857d-0f85d286cdb9" containerName="extract" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.502664 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.504790 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8c7pc" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.505064 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.505378 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.511866 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488"] Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.529373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psn9l\" (UniqueName: \"kubernetes.io/projected/f0b0d865-5ec1-4dff-ada5-658bd9676d32-kube-api-access-psn9l\") pod \"nmstate-operator-5b5b58f5c8-bz488\" (UID: \"f0b0d865-5ec1-4dff-ada5-658bd9676d32\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.630276 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psn9l\" (UniqueName: \"kubernetes.io/projected/f0b0d865-5ec1-4dff-ada5-658bd9676d32-kube-api-access-psn9l\") pod \"nmstate-operator-5b5b58f5c8-bz488\" (UID: \"f0b0d865-5ec1-4dff-ada5-658bd9676d32\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.654888 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psn9l\" (UniqueName: \"kubernetes.io/projected/f0b0d865-5ec1-4dff-ada5-658bd9676d32-kube-api-access-psn9l\") pod \"nmstate-operator-5b5b58f5c8-bz488\" (UID: \"f0b0d865-5ec1-4dff-ada5-658bd9676d32\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" Dec 04 06:23:32 crc kubenswrapper[4685]: I1204 06:23:32.817768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" Dec 04 06:23:33 crc kubenswrapper[4685]: I1204 06:23:33.316315 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488"] Dec 04 06:23:34 crc kubenswrapper[4685]: I1204 06:23:34.186120 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" event={"ID":"f0b0d865-5ec1-4dff-ada5-658bd9676d32","Type":"ContainerStarted","Data":"dc55876d96a3573cdaeff0402580dabb4ae3b3f6266a0299318ae66993c3843e"} Dec 04 06:23:36 crc kubenswrapper[4685]: I1204 06:23:36.121942 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:23:36 crc kubenswrapper[4685]: I1204 06:23:36.122037 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:23:36 crc kubenswrapper[4685]: I1204 06:23:36.208068 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" event={"ID":"f0b0d865-5ec1-4dff-ada5-658bd9676d32","Type":"ContainerStarted","Data":"50936d65d461681111b3e8939f802e9bc01887faaaf06a0090ed15c9effa37be"} Dec 04 06:23:36 crc kubenswrapper[4685]: I1204 06:23:36.239449 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bz488" podStartSLOduration=2.187811534 podStartE2EDuration="4.239402542s" podCreationTimestamp="2025-12-04 06:23:32 +0000 UTC" firstStartedPulling="2025-12-04 06:23:33.318746232 +0000 UTC m=+750.506977047" lastFinishedPulling="2025-12-04 06:23:35.37033728 +0000 UTC m=+752.558568055" observedRunningTime="2025-12-04 06:23:36.230872539 +0000 UTC m=+753.419103384" watchObservedRunningTime="2025-12-04 06:23:36.239402542 +0000 UTC m=+753.427633357" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.233622 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.234627 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.240882 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-b7v7w" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.241889 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.244326 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.246318 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.276012 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.283962 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.294439 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-wpz9x"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.295971 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.374822 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.375482 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.377963 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.377980 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.378497 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-4gjmt" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.391585 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400274 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-dbus-socket\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400346 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ec56bcce-1006-4645-a67d-2a4405768a11-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9hhls\" (UID: \"ec56bcce-1006-4645-a67d-2a4405768a11\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400374 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqss2\" (UniqueName: \"kubernetes.io/projected/a3048137-1da6-4838-99bf-f4ab08bf895c-kube-api-access-gqss2\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400401 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhlbz\" (UniqueName: \"kubernetes.io/projected/ec56bcce-1006-4645-a67d-2a4405768a11-kube-api-access-lhlbz\") pod \"nmstate-webhook-5f6d4c5ccb-9hhls\" (UID: \"ec56bcce-1006-4645-a67d-2a4405768a11\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400453 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-nmstate-lock\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400499 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vtx\" (UniqueName: \"kubernetes.io/projected/bef52eb9-45f0-43e5-b675-e4d3e819a3ec-kube-api-access-f7vtx\") pod \"nmstate-metrics-7f946cbc9-t55gj\" (UID: \"bef52eb9-45f0-43e5-b675-e4d3e819a3ec\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.400518 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-ovs-socket\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.501828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vtx\" (UniqueName: \"kubernetes.io/projected/bef52eb9-45f0-43e5-b675-e4d3e819a3ec-kube-api-access-f7vtx\") pod \"nmstate-metrics-7f946cbc9-t55gj\" (UID: \"bef52eb9-45f0-43e5-b675-e4d3e819a3ec\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.501887 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-ovs-socket\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.501929 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-dbus-socket\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502060 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502110 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ec56bcce-1006-4645-a67d-2a4405768a11-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9hhls\" (UID: \"ec56bcce-1006-4645-a67d-2a4405768a11\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502153 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqss2\" (UniqueName: \"kubernetes.io/projected/a3048137-1da6-4838-99bf-f4ab08bf895c-kube-api-access-gqss2\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502184 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502207 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhlbz\" (UniqueName: \"kubernetes.io/projected/ec56bcce-1006-4645-a67d-2a4405768a11-kube-api-access-lhlbz\") pod \"nmstate-webhook-5f6d4c5ccb-9hhls\" (UID: \"ec56bcce-1006-4645-a67d-2a4405768a11\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502244 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rhh8\" (UniqueName: \"kubernetes.io/projected/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-kube-api-access-7rhh8\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502283 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-nmstate-lock\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502357 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-nmstate-lock\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.502697 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-ovs-socket\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.503119 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a3048137-1da6-4838-99bf-f4ab08bf895c-dbus-socket\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.512219 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ec56bcce-1006-4645-a67d-2a4405768a11-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9hhls\" (UID: \"ec56bcce-1006-4645-a67d-2a4405768a11\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.524027 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vtx\" (UniqueName: \"kubernetes.io/projected/bef52eb9-45f0-43e5-b675-e4d3e819a3ec-kube-api-access-f7vtx\") pod \"nmstate-metrics-7f946cbc9-t55gj\" (UID: \"bef52eb9-45f0-43e5-b675-e4d3e819a3ec\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.528324 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhlbz\" (UniqueName: \"kubernetes.io/projected/ec56bcce-1006-4645-a67d-2a4405768a11-kube-api-access-lhlbz\") pod \"nmstate-webhook-5f6d4c5ccb-9hhls\" (UID: \"ec56bcce-1006-4645-a67d-2a4405768a11\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.531909 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqss2\" (UniqueName: \"kubernetes.io/projected/a3048137-1da6-4838-99bf-f4ab08bf895c-kube-api-access-gqss2\") pod \"nmstate-handler-wpz9x\" (UID: \"a3048137-1da6-4838-99bf-f4ab08bf895c\") " pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.559592 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.570189 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5ffb5f4cbf-trglj"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.571591 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.590967 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.596525 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5ffb5f4cbf-trglj"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.603899 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.603966 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rhh8\" (UniqueName: \"kubernetes.io/projected/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-kube-api-access-7rhh8\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.604049 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.606032 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.615564 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.623159 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.649522 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rhh8\" (UniqueName: \"kubernetes.io/projected/f77ee3f7-2978-4a54-ac7e-11a038bd3f5e-kube-api-access-7rhh8\") pod \"nmstate-console-plugin-7fbb5f6569-rnj2j\" (UID: \"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.694929 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.706758 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-serving-cert\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.706900 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-trusted-ca-bundle\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.706945 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-service-ca\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.706969 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5g66\" (UniqueName: \"kubernetes.io/projected/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-kube-api-access-s5g66\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.707024 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-config\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.707219 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-oauth-config\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.707344 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-oauth-serving-cert\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.809658 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-service-ca\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.810246 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5g66\" (UniqueName: \"kubernetes.io/projected/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-kube-api-access-s5g66\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.810314 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-config\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.810342 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-oauth-config\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.810397 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-oauth-serving-cert\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.810451 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-serving-cert\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.810497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-trusted-ca-bundle\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.812592 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-config\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.812804 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-service-ca\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.813307 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-oauth-serving-cert\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.813957 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-trusted-ca-bundle\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.821057 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-oauth-config\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.825657 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj"] Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.827846 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-console-serving-cert\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.838948 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5g66\" (UniqueName: \"kubernetes.io/projected/6ac4533a-86eb-4505-82a1-6ffdbf17ed72-kube-api-access-s5g66\") pod \"console-5ffb5f4cbf-trglj\" (UID: \"6ac4533a-86eb-4505-82a1-6ffdbf17ed72\") " pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:37 crc kubenswrapper[4685]: W1204 06:23:37.842158 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbef52eb9_45f0_43e5_b675_e4d3e819a3ec.slice/crio-2fef44b8838d9b26b504cf53fcf4249e098e02a0d110688f0956cb77d6aeaa40 WatchSource:0}: Error finding container 2fef44b8838d9b26b504cf53fcf4249e098e02a0d110688f0956cb77d6aeaa40: Status 404 returned error can't find the container with id 2fef44b8838d9b26b504cf53fcf4249e098e02a0d110688f0956cb77d6aeaa40 Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.878517 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls"] Dec 04 06:23:37 crc kubenswrapper[4685]: W1204 06:23:37.887625 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec56bcce_1006_4645_a67d_2a4405768a11.slice/crio-f0a8871bc53489d1eee59036455ce45642f4fb31b0543020af7b90b3dda43e75 WatchSource:0}: Error finding container f0a8871bc53489d1eee59036455ce45642f4fb31b0543020af7b90b3dda43e75: Status 404 returned error can't find the container with id f0a8871bc53489d1eee59036455ce45642f4fb31b0543020af7b90b3dda43e75 Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.934729 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j"] Dec 04 06:23:37 crc kubenswrapper[4685]: W1204 06:23:37.939847 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf77ee3f7_2978_4a54_ac7e_11a038bd3f5e.slice/crio-cc21d5b7467c67503de3851b7e4cb47164497c3de27585964d52f822b10455ac WatchSource:0}: Error finding container cc21d5b7467c67503de3851b7e4cb47164497c3de27585964d52f822b10455ac: Status 404 returned error can't find the container with id cc21d5b7467c67503de3851b7e4cb47164497c3de27585964d52f822b10455ac Dec 04 06:23:37 crc kubenswrapper[4685]: I1204 06:23:37.945658 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:38 crc kubenswrapper[4685]: I1204 06:23:38.145247 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5ffb5f4cbf-trglj"] Dec 04 06:23:38 crc kubenswrapper[4685]: W1204 06:23:38.153264 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ac4533a_86eb_4505_82a1_6ffdbf17ed72.slice/crio-eedad53cb1353309ce645d94e212fe442abd29f754d130adbc11c7921a6cb45a WatchSource:0}: Error finding container eedad53cb1353309ce645d94e212fe442abd29f754d130adbc11c7921a6cb45a: Status 404 returned error can't find the container with id eedad53cb1353309ce645d94e212fe442abd29f754d130adbc11c7921a6cb45a Dec 04 06:23:38 crc kubenswrapper[4685]: I1204 06:23:38.227824 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" event={"ID":"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e","Type":"ContainerStarted","Data":"cc21d5b7467c67503de3851b7e4cb47164497c3de27585964d52f822b10455ac"} Dec 04 06:23:38 crc kubenswrapper[4685]: I1204 06:23:38.230042 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" event={"ID":"ec56bcce-1006-4645-a67d-2a4405768a11","Type":"ContainerStarted","Data":"f0a8871bc53489d1eee59036455ce45642f4fb31b0543020af7b90b3dda43e75"} Dec 04 06:23:38 crc kubenswrapper[4685]: I1204 06:23:38.231189 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" event={"ID":"bef52eb9-45f0-43e5-b675-e4d3e819a3ec","Type":"ContainerStarted","Data":"2fef44b8838d9b26b504cf53fcf4249e098e02a0d110688f0956cb77d6aeaa40"} Dec 04 06:23:38 crc kubenswrapper[4685]: I1204 06:23:38.233580 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wpz9x" event={"ID":"a3048137-1da6-4838-99bf-f4ab08bf895c","Type":"ContainerStarted","Data":"c8a3d35d638085a4f9c75cd040ab56c0d38b013039e54ac0741b812b624e2516"} Dec 04 06:23:38 crc kubenswrapper[4685]: I1204 06:23:38.234932 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5ffb5f4cbf-trglj" event={"ID":"6ac4533a-86eb-4505-82a1-6ffdbf17ed72","Type":"ContainerStarted","Data":"eedad53cb1353309ce645d94e212fe442abd29f754d130adbc11c7921a6cb45a"} Dec 04 06:23:39 crc kubenswrapper[4685]: I1204 06:23:39.244632 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5ffb5f4cbf-trglj" event={"ID":"6ac4533a-86eb-4505-82a1-6ffdbf17ed72","Type":"ContainerStarted","Data":"5641ffa72c49b9270764e1ccb0329f7b402a7dfef5251445b432ac535afedad5"} Dec 04 06:23:39 crc kubenswrapper[4685]: I1204 06:23:39.262847 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5ffb5f4cbf-trglj" podStartSLOduration=2.2628309189999998 podStartE2EDuration="2.262830919s" podCreationTimestamp="2025-12-04 06:23:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:23:39.259095503 +0000 UTC m=+756.447326318" watchObservedRunningTime="2025-12-04 06:23:39.262830919 +0000 UTC m=+756.451061694" Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.259680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" event={"ID":"bef52eb9-45f0-43e5-b675-e4d3e819a3ec","Type":"ContainerStarted","Data":"b2892eb099c50f388d1e71afbc4f49ae149dc86f5e28ad64210888676251e4cd"} Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.262473 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wpz9x" event={"ID":"a3048137-1da6-4838-99bf-f4ab08bf895c","Type":"ContainerStarted","Data":"9919d92dda60aa5569eaa118131e0aba86dd3882996d80f4ea21efa04dae506e"} Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.263438 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.274466 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" event={"ID":"f77ee3f7-2978-4a54-ac7e-11a038bd3f5e","Type":"ContainerStarted","Data":"c3b42f7c406f3dea09ec9ae28eab40e53894311f21c4e428ad717fbb6b8f052d"} Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.277313 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" event={"ID":"ec56bcce-1006-4645-a67d-2a4405768a11","Type":"ContainerStarted","Data":"c3b4bea984a22d8e6477767210bb4e8713f77f2b3d7433cd9e374979a86a95a1"} Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.277753 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.297352 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-wpz9x" podStartSLOduration=1.267129831 podStartE2EDuration="4.297323401s" podCreationTimestamp="2025-12-04 06:23:37 +0000 UTC" firstStartedPulling="2025-12-04 06:23:37.651782474 +0000 UTC m=+754.840013249" lastFinishedPulling="2025-12-04 06:23:40.681976034 +0000 UTC m=+757.870206819" observedRunningTime="2025-12-04 06:23:41.290581918 +0000 UTC m=+758.478812713" watchObservedRunningTime="2025-12-04 06:23:41.297323401 +0000 UTC m=+758.485554196" Dec 04 06:23:41 crc kubenswrapper[4685]: I1204 06:23:41.335145 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" podStartSLOduration=1.532963472 podStartE2EDuration="4.33512726s" podCreationTimestamp="2025-12-04 06:23:37 +0000 UTC" firstStartedPulling="2025-12-04 06:23:37.888884836 +0000 UTC m=+755.077115611" lastFinishedPulling="2025-12-04 06:23:40.691048614 +0000 UTC m=+757.879279399" observedRunningTime="2025-12-04 06:23:41.329764437 +0000 UTC m=+758.517995212" watchObservedRunningTime="2025-12-04 06:23:41.33512726 +0000 UTC m=+758.523358045" Dec 04 06:23:43 crc kubenswrapper[4685]: I1204 06:23:43.166224 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rnj2j" podStartSLOduration=3.447091984 podStartE2EDuration="6.166182669s" podCreationTimestamp="2025-12-04 06:23:37 +0000 UTC" firstStartedPulling="2025-12-04 06:23:37.944434803 +0000 UTC m=+755.132665578" lastFinishedPulling="2025-12-04 06:23:40.663525478 +0000 UTC m=+757.851756263" observedRunningTime="2025-12-04 06:23:41.355525203 +0000 UTC m=+758.543755988" watchObservedRunningTime="2025-12-04 06:23:43.166182669 +0000 UTC m=+760.354413474" Dec 04 06:23:43 crc kubenswrapper[4685]: I1204 06:23:43.297557 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" event={"ID":"bef52eb9-45f0-43e5-b675-e4d3e819a3ec","Type":"ContainerStarted","Data":"ee6a2ec1736650b192805ab8ee4ef427038cce02da4d2b129f8b5554c255c910"} Dec 04 06:23:43 crc kubenswrapper[4685]: I1204 06:23:43.324391 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-t55gj" podStartSLOduration=1.337413698 podStartE2EDuration="6.324359068s" podCreationTimestamp="2025-12-04 06:23:37 +0000 UTC" firstStartedPulling="2025-12-04 06:23:37.846369862 +0000 UTC m=+755.034600637" lastFinishedPulling="2025-12-04 06:23:42.833315232 +0000 UTC m=+760.021546007" observedRunningTime="2025-12-04 06:23:43.323481702 +0000 UTC m=+760.511712547" watchObservedRunningTime="2025-12-04 06:23:43.324359068 +0000 UTC m=+760.512589883" Dec 04 06:23:47 crc kubenswrapper[4685]: I1204 06:23:47.652746 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-wpz9x" Dec 04 06:23:47 crc kubenswrapper[4685]: I1204 06:23:47.947149 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:47 crc kubenswrapper[4685]: I1204 06:23:47.947233 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:47 crc kubenswrapper[4685]: I1204 06:23:47.954861 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:48 crc kubenswrapper[4685]: I1204 06:23:48.055288 4685 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 06:23:48 crc kubenswrapper[4685]: I1204 06:23:48.340464 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5ffb5f4cbf-trglj" Dec 04 06:23:48 crc kubenswrapper[4685]: I1204 06:23:48.400308 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s6kbn"] Dec 04 06:23:57 crc kubenswrapper[4685]: I1204 06:23:57.599148 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9hhls" Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.121747 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.122455 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.122522 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.123220 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39d6b27b9ba56cb920b538ba51d04d90d1b790d65f04ecd9199b893b79b8e599"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.123311 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://39d6b27b9ba56cb920b538ba51d04d90d1b790d65f04ecd9199b893b79b8e599" gracePeriod=600 Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.500950 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="39d6b27b9ba56cb920b538ba51d04d90d1b790d65f04ecd9199b893b79b8e599" exitCode=0 Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.501105 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"39d6b27b9ba56cb920b538ba51d04d90d1b790d65f04ecd9199b893b79b8e599"} Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.501787 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11"} Dec 04 06:24:06 crc kubenswrapper[4685]: I1204 06:24:06.501829 4685 scope.go:117] "RemoveContainer" containerID="2dda32d5375f2205695b926293374a6c3db5c6e9ff6115114489ad170e81d7dc" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.447834 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-s6kbn" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" containerName="console" containerID="cri-o://f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31" gracePeriod=15 Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.870305 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s6kbn_534c7808-899b-4952-9a29-b65dacb98ff0/console/0.log" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.870404 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943219 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-trusted-ca-bundle\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943315 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-oauth-serving-cert\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943365 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-oauth-config\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943459 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-serving-cert\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943532 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-console-config\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943576 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n59pf\" (UniqueName: \"kubernetes.io/projected/534c7808-899b-4952-9a29-b65dacb98ff0-kube-api-access-n59pf\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.943640 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-service-ca\") pod \"534c7808-899b-4952-9a29-b65dacb98ff0\" (UID: \"534c7808-899b-4952-9a29-b65dacb98ff0\") " Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.944634 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-console-config" (OuterVolumeSpecName: "console-config") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.944793 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.944808 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.944917 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-service-ca" (OuterVolumeSpecName: "service-ca") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.950840 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534c7808-899b-4952-9a29-b65dacb98ff0-kube-api-access-n59pf" (OuterVolumeSpecName: "kube-api-access-n59pf") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "kube-api-access-n59pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.951390 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:24:13 crc kubenswrapper[4685]: I1204 06:24:13.951953 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "534c7808-899b-4952-9a29-b65dacb98ff0" (UID: "534c7808-899b-4952-9a29-b65dacb98ff0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045019 4685 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045073 4685 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045103 4685 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/534c7808-899b-4952-9a29-b65dacb98ff0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045129 4685 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045153 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n59pf\" (UniqueName: \"kubernetes.io/projected/534c7808-899b-4952-9a29-b65dacb98ff0-kube-api-access-n59pf\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045180 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.045203 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/534c7808-899b-4952-9a29-b65dacb98ff0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.182309 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z92mv"] Dec 04 06:24:14 crc kubenswrapper[4685]: E1204 06:24:14.186301 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" containerName="console" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.186329 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" containerName="console" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.186621 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" containerName="console" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.188903 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.198603 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z92mv"] Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.248223 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-catalog-content\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.248325 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgc7r\" (UniqueName: \"kubernetes.io/projected/c018a366-2f44-45a5-ba15-23f4b9d8cb64-kube-api-access-lgc7r\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.248382 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-utilities\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.349862 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgc7r\" (UniqueName: \"kubernetes.io/projected/c018a366-2f44-45a5-ba15-23f4b9d8cb64-kube-api-access-lgc7r\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.349935 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-utilities\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.349993 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-catalog-content\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.350596 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-utilities\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.350641 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-catalog-content\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.378133 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgc7r\" (UniqueName: \"kubernetes.io/projected/c018a366-2f44-45a5-ba15-23f4b9d8cb64-kube-api-access-lgc7r\") pod \"redhat-marketplace-z92mv\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.523607 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.570664 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s6kbn_534c7808-899b-4952-9a29-b65dacb98ff0/console/0.log" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.570720 4685 generic.go:334] "Generic (PLEG): container finished" podID="534c7808-899b-4952-9a29-b65dacb98ff0" containerID="f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31" exitCode=2 Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.570752 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6kbn" event={"ID":"534c7808-899b-4952-9a29-b65dacb98ff0","Type":"ContainerDied","Data":"f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31"} Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.570787 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6kbn" event={"ID":"534c7808-899b-4952-9a29-b65dacb98ff0","Type":"ContainerDied","Data":"dbaf752c85d44102e4754c1948f6d24947e2e8c533c2373f40d0148de00348a2"} Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.570807 4685 scope.go:117] "RemoveContainer" containerID="f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.570839 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6kbn" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.612180 4685 scope.go:117] "RemoveContainer" containerID="f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31" Dec 04 06:24:14 crc kubenswrapper[4685]: E1204 06:24:14.613077 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31\": container with ID starting with f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31 not found: ID does not exist" containerID="f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.613163 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31"} err="failed to get container status \"f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31\": rpc error: code = NotFound desc = could not find container \"f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31\": container with ID starting with f2230deefac8df31ecc5a6ff459dfb55be5e23ef7ab9cce896a72ef22f0b5f31 not found: ID does not exist" Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.621917 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s6kbn"] Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.623073 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-s6kbn"] Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.794261 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z92mv"] Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.815912 4685 patch_prober.go:28] interesting pod/console-f9d7485db-s6kbn container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 06:24:14 crc kubenswrapper[4685]: I1204 06:24:14.815999 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-s6kbn" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 06:24:15 crc kubenswrapper[4685]: I1204 06:24:15.137484 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534c7808-899b-4952-9a29-b65dacb98ff0" path="/var/lib/kubelet/pods/534c7808-899b-4952-9a29-b65dacb98ff0/volumes" Dec 04 06:24:15 crc kubenswrapper[4685]: I1204 06:24:15.579752 4685 generic.go:334] "Generic (PLEG): container finished" podID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerID="1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68" exitCode=0 Dec 04 06:24:15 crc kubenswrapper[4685]: I1204 06:24:15.579886 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z92mv" event={"ID":"c018a366-2f44-45a5-ba15-23f4b9d8cb64","Type":"ContainerDied","Data":"1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68"} Dec 04 06:24:15 crc kubenswrapper[4685]: I1204 06:24:15.579923 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z92mv" event={"ID":"c018a366-2f44-45a5-ba15-23f4b9d8cb64","Type":"ContainerStarted","Data":"d751c59a5e4f3697a07f6aea83357122f030523157ea041c77b475838f3876c5"} Dec 04 06:24:15 crc kubenswrapper[4685]: I1204 06:24:15.996224 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl"] Dec 04 06:24:15 crc kubenswrapper[4685]: I1204 06:24:15.999008 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.002641 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.010900 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl"] Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.072503 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxh84\" (UniqueName: \"kubernetes.io/projected/46d378f4-3618-4d17-a705-dc3b42de15bc-kube-api-access-cxh84\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.072578 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.072652 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.173714 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxh84\" (UniqueName: \"kubernetes.io/projected/46d378f4-3618-4d17-a705-dc3b42de15bc-kube-api-access-cxh84\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.173793 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.173919 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.174787 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.174996 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.204000 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxh84\" (UniqueName: \"kubernetes.io/projected/46d378f4-3618-4d17-a705-dc3b42de15bc-kube-api-access-cxh84\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.321019 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.591445 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z92mv" event={"ID":"c018a366-2f44-45a5-ba15-23f4b9d8cb64","Type":"ContainerDied","Data":"4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557"} Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.591388 4685 generic.go:334] "Generic (PLEG): container finished" podID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerID="4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557" exitCode=0 Dec 04 06:24:16 crc kubenswrapper[4685]: I1204 06:24:16.824828 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl"] Dec 04 06:24:16 crc kubenswrapper[4685]: W1204 06:24:16.837901 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46d378f4_3618_4d17_a705_dc3b42de15bc.slice/crio-ba594bb6695be5196bb5a76d4db09f8cc409f3fcc223d1cd06262b1e582c0cad WatchSource:0}: Error finding container ba594bb6695be5196bb5a76d4db09f8cc409f3fcc223d1cd06262b1e582c0cad: Status 404 returned error can't find the container with id ba594bb6695be5196bb5a76d4db09f8cc409f3fcc223d1cd06262b1e582c0cad Dec 04 06:24:17 crc kubenswrapper[4685]: I1204 06:24:17.603243 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z92mv" event={"ID":"c018a366-2f44-45a5-ba15-23f4b9d8cb64","Type":"ContainerStarted","Data":"8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d"} Dec 04 06:24:17 crc kubenswrapper[4685]: I1204 06:24:17.605643 4685 generic.go:334] "Generic (PLEG): container finished" podID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerID="28d3b728bd990dc0e99acc25350cc08344c89c4ac9bd78532825f6123485fcfd" exitCode=0 Dec 04 06:24:17 crc kubenswrapper[4685]: I1204 06:24:17.605709 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" event={"ID":"46d378f4-3618-4d17-a705-dc3b42de15bc","Type":"ContainerDied","Data":"28d3b728bd990dc0e99acc25350cc08344c89c4ac9bd78532825f6123485fcfd"} Dec 04 06:24:17 crc kubenswrapper[4685]: I1204 06:24:17.605755 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" event={"ID":"46d378f4-3618-4d17-a705-dc3b42de15bc","Type":"ContainerStarted","Data":"ba594bb6695be5196bb5a76d4db09f8cc409f3fcc223d1cd06262b1e582c0cad"} Dec 04 06:24:17 crc kubenswrapper[4685]: I1204 06:24:17.643749 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z92mv" podStartSLOduration=2.202020642 podStartE2EDuration="3.643698577s" podCreationTimestamp="2025-12-04 06:24:14 +0000 UTC" firstStartedPulling="2025-12-04 06:24:15.582217955 +0000 UTC m=+792.770448740" lastFinishedPulling="2025-12-04 06:24:17.0238959 +0000 UTC m=+794.212126675" observedRunningTime="2025-12-04 06:24:17.636392063 +0000 UTC m=+794.824622858" watchObservedRunningTime="2025-12-04 06:24:17.643698577 +0000 UTC m=+794.831929382" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.558391 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qcfh4"] Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.561470 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.568763 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcfh4"] Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.622451 4685 generic.go:334] "Generic (PLEG): container finished" podID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerID="d71b7856f7d83201ac452b7e8e73a0488b143e93a35adaef2e7d5bac5f1a7b48" exitCode=0 Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.622515 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" event={"ID":"46d378f4-3618-4d17-a705-dc3b42de15bc","Type":"ContainerDied","Data":"d71b7856f7d83201ac452b7e8e73a0488b143e93a35adaef2e7d5bac5f1a7b48"} Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.623383 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-catalog-content\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.623551 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9vq7\" (UniqueName: \"kubernetes.io/projected/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-kube-api-access-h9vq7\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.623761 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-utilities\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.724982 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-utilities\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.725138 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-catalog-content\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.725194 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9vq7\" (UniqueName: \"kubernetes.io/projected/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-kube-api-access-h9vq7\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.725786 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-utilities\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.725971 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-catalog-content\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.760330 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9vq7\" (UniqueName: \"kubernetes.io/projected/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-kube-api-access-h9vq7\") pod \"redhat-operators-qcfh4\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:19 crc kubenswrapper[4685]: I1204 06:24:19.892978 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:20 crc kubenswrapper[4685]: I1204 06:24:20.146818 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcfh4"] Dec 04 06:24:20 crc kubenswrapper[4685]: I1204 06:24:20.633165 4685 generic.go:334] "Generic (PLEG): container finished" podID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerID="721a67eb14ce81cefa0e807745c1e3eed69cbee70a0c2b36660f57a9ee3a11ae" exitCode=0 Dec 04 06:24:20 crc kubenswrapper[4685]: I1204 06:24:20.633301 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerDied","Data":"721a67eb14ce81cefa0e807745c1e3eed69cbee70a0c2b36660f57a9ee3a11ae"} Dec 04 06:24:20 crc kubenswrapper[4685]: I1204 06:24:20.633354 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerStarted","Data":"557ae59bf971e12f4d3e5fd5328fca7d8483422c6f0c7f051de1c6c634ee1b98"} Dec 04 06:24:20 crc kubenswrapper[4685]: I1204 06:24:20.638762 4685 generic.go:334] "Generic (PLEG): container finished" podID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerID="e632e601c854da9f2793145bf70ea7b02a3f5480fcc988e53606a036804a3a3a" exitCode=0 Dec 04 06:24:20 crc kubenswrapper[4685]: I1204 06:24:20.638823 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" event={"ID":"46d378f4-3618-4d17-a705-dc3b42de15bc","Type":"ContainerDied","Data":"e632e601c854da9f2793145bf70ea7b02a3f5480fcc988e53606a036804a3a3a"} Dec 04 06:24:21 crc kubenswrapper[4685]: I1204 06:24:21.660995 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerStarted","Data":"5419af0a92cb9c6171193d6f6a8cee23a77f4970a1c3f7f7a8ccecdebc264854"} Dec 04 06:24:21 crc kubenswrapper[4685]: I1204 06:24:21.991776 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.065257 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-util\") pod \"46d378f4-3618-4d17-a705-dc3b42de15bc\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.065294 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-bundle\") pod \"46d378f4-3618-4d17-a705-dc3b42de15bc\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.065399 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxh84\" (UniqueName: \"kubernetes.io/projected/46d378f4-3618-4d17-a705-dc3b42de15bc-kube-api-access-cxh84\") pod \"46d378f4-3618-4d17-a705-dc3b42de15bc\" (UID: \"46d378f4-3618-4d17-a705-dc3b42de15bc\") " Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.066446 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-bundle" (OuterVolumeSpecName: "bundle") pod "46d378f4-3618-4d17-a705-dc3b42de15bc" (UID: "46d378f4-3618-4d17-a705-dc3b42de15bc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.071568 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46d378f4-3618-4d17-a705-dc3b42de15bc-kube-api-access-cxh84" (OuterVolumeSpecName: "kube-api-access-cxh84") pod "46d378f4-3618-4d17-a705-dc3b42de15bc" (UID: "46d378f4-3618-4d17-a705-dc3b42de15bc"). InnerVolumeSpecName "kube-api-access-cxh84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.078453 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-util" (OuterVolumeSpecName: "util") pod "46d378f4-3618-4d17-a705-dc3b42de15bc" (UID: "46d378f4-3618-4d17-a705-dc3b42de15bc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.167252 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxh84\" (UniqueName: \"kubernetes.io/projected/46d378f4-3618-4d17-a705-dc3b42de15bc-kube-api-access-cxh84\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.167286 4685 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-util\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.167295 4685 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46d378f4-3618-4d17-a705-dc3b42de15bc-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.671233 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.671258 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl" event={"ID":"46d378f4-3618-4d17-a705-dc3b42de15bc","Type":"ContainerDied","Data":"ba594bb6695be5196bb5a76d4db09f8cc409f3fcc223d1cd06262b1e582c0cad"} Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.672078 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba594bb6695be5196bb5a76d4db09f8cc409f3fcc223d1cd06262b1e582c0cad" Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.674642 4685 generic.go:334] "Generic (PLEG): container finished" podID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerID="5419af0a92cb9c6171193d6f6a8cee23a77f4970a1c3f7f7a8ccecdebc264854" exitCode=0 Dec 04 06:24:22 crc kubenswrapper[4685]: I1204 06:24:22.674692 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerDied","Data":"5419af0a92cb9c6171193d6f6a8cee23a77f4970a1c3f7f7a8ccecdebc264854"} Dec 04 06:24:23 crc kubenswrapper[4685]: I1204 06:24:23.685841 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerStarted","Data":"20d25ab279a3efe58a23f67c5d9c2cd459daf78f92b780250b16f04b8c03692c"} Dec 04 06:24:23 crc kubenswrapper[4685]: I1204 06:24:23.707133 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qcfh4" podStartSLOduration=2.252619109 podStartE2EDuration="4.707109129s" podCreationTimestamp="2025-12-04 06:24:19 +0000 UTC" firstStartedPulling="2025-12-04 06:24:20.635114986 +0000 UTC m=+797.823345791" lastFinishedPulling="2025-12-04 06:24:23.089604996 +0000 UTC m=+800.277835811" observedRunningTime="2025-12-04 06:24:23.706563781 +0000 UTC m=+800.894794566" watchObservedRunningTime="2025-12-04 06:24:23.707109129 +0000 UTC m=+800.895339904" Dec 04 06:24:24 crc kubenswrapper[4685]: I1204 06:24:24.524354 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:24 crc kubenswrapper[4685]: I1204 06:24:24.524423 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:24 crc kubenswrapper[4685]: I1204 06:24:24.575050 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:24 crc kubenswrapper[4685]: I1204 06:24:24.725686 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:27 crc kubenswrapper[4685]: I1204 06:24:27.732157 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z92mv"] Dec 04 06:24:27 crc kubenswrapper[4685]: I1204 06:24:27.732730 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z92mv" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="registry-server" containerID="cri-o://8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d" gracePeriod=2 Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.668830 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.714331 4685 generic.go:334] "Generic (PLEG): container finished" podID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerID="8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d" exitCode=0 Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.714370 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z92mv" event={"ID":"c018a366-2f44-45a5-ba15-23f4b9d8cb64","Type":"ContainerDied","Data":"8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d"} Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.714398 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z92mv" event={"ID":"c018a366-2f44-45a5-ba15-23f4b9d8cb64","Type":"ContainerDied","Data":"d751c59a5e4f3697a07f6aea83357122f030523157ea041c77b475838f3876c5"} Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.714431 4685 scope.go:117] "RemoveContainer" containerID="8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.714537 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z92mv" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.737010 4685 scope.go:117] "RemoveContainer" containerID="4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.759481 4685 scope.go:117] "RemoveContainer" containerID="1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.764605 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgc7r\" (UniqueName: \"kubernetes.io/projected/c018a366-2f44-45a5-ba15-23f4b9d8cb64-kube-api-access-lgc7r\") pod \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.764669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-catalog-content\") pod \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.764808 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-utilities\") pod \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\" (UID: \"c018a366-2f44-45a5-ba15-23f4b9d8cb64\") " Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.765720 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-utilities" (OuterVolumeSpecName: "utilities") pod "c018a366-2f44-45a5-ba15-23f4b9d8cb64" (UID: "c018a366-2f44-45a5-ba15-23f4b9d8cb64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.775621 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c018a366-2f44-45a5-ba15-23f4b9d8cb64-kube-api-access-lgc7r" (OuterVolumeSpecName: "kube-api-access-lgc7r") pod "c018a366-2f44-45a5-ba15-23f4b9d8cb64" (UID: "c018a366-2f44-45a5-ba15-23f4b9d8cb64"). InnerVolumeSpecName "kube-api-access-lgc7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.790286 4685 scope.go:117] "RemoveContainer" containerID="8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d" Dec 04 06:24:28 crc kubenswrapper[4685]: E1204 06:24:28.792517 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d\": container with ID starting with 8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d not found: ID does not exist" containerID="8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.792558 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d"} err="failed to get container status \"8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d\": rpc error: code = NotFound desc = could not find container \"8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d\": container with ID starting with 8107fd4df8783d29e52d8fa01ce02fa449eef0acab0f196e373c23033b0a9a0d not found: ID does not exist" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.792584 4685 scope.go:117] "RemoveContainer" containerID="4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.792602 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c018a366-2f44-45a5-ba15-23f4b9d8cb64" (UID: "c018a366-2f44-45a5-ba15-23f4b9d8cb64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:24:28 crc kubenswrapper[4685]: E1204 06:24:28.792931 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557\": container with ID starting with 4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557 not found: ID does not exist" containerID="4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.793022 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557"} err="failed to get container status \"4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557\": rpc error: code = NotFound desc = could not find container \"4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557\": container with ID starting with 4a964e44707654a67b440693e0615996e3c9f1c6e0afc463cb2fdc44b80f5557 not found: ID does not exist" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.793104 4685 scope.go:117] "RemoveContainer" containerID="1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68" Dec 04 06:24:28 crc kubenswrapper[4685]: E1204 06:24:28.793309 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68\": container with ID starting with 1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68 not found: ID does not exist" containerID="1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.793394 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68"} err="failed to get container status \"1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68\": rpc error: code = NotFound desc = could not find container \"1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68\": container with ID starting with 1745db6e5c2cb83c69fc84fad03ae520e609d9299dd2c7c319b156990b1d6f68 not found: ID does not exist" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.866059 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.866095 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgc7r\" (UniqueName: \"kubernetes.io/projected/c018a366-2f44-45a5-ba15-23f4b9d8cb64-kube-api-access-lgc7r\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:28 crc kubenswrapper[4685]: I1204 06:24:28.866108 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c018a366-2f44-45a5-ba15-23f4b9d8cb64-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.048510 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z92mv"] Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.072830 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z92mv"] Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.132372 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" path="/var/lib/kubelet/pods/c018a366-2f44-45a5-ba15-23f4b9d8cb64/volumes" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.893665 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.894927 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.940364 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971524 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57"] Dec 04 06:24:29 crc kubenswrapper[4685]: E1204 06:24:29.971778 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="util" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971797 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="util" Dec 04 06:24:29 crc kubenswrapper[4685]: E1204 06:24:29.971825 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="extract-utilities" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971834 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="extract-utilities" Dec 04 06:24:29 crc kubenswrapper[4685]: E1204 06:24:29.971848 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="extract" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971856 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="extract" Dec 04 06:24:29 crc kubenswrapper[4685]: E1204 06:24:29.971868 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="pull" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971877 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="pull" Dec 04 06:24:29 crc kubenswrapper[4685]: E1204 06:24:29.971891 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="extract-content" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971900 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="extract-content" Dec 04 06:24:29 crc kubenswrapper[4685]: E1204 06:24:29.971914 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="registry-server" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.971921 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="registry-server" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.972060 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d378f4-3618-4d17-a705-dc3b42de15bc" containerName="extract" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.972084 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c018a366-2f44-45a5-ba15-23f4b9d8cb64" containerName="registry-server" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.972546 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.974488 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-f8mtq" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.974769 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.975325 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.975505 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.975500 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 06:24:29 crc kubenswrapper[4685]: I1204 06:24:29.993249 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57"] Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.079319 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-webhook-cert\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.079385 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gp72\" (UniqueName: \"kubernetes.io/projected/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-kube-api-access-6gp72\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.079461 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-apiservice-cert\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.181113 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-webhook-cert\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.181184 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gp72\" (UniqueName: \"kubernetes.io/projected/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-kube-api-access-6gp72\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.181220 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-apiservice-cert\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.190663 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-apiservice-cert\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.192581 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-webhook-cert\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.202435 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8"] Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.203338 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.211058 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gp72\" (UniqueName: \"kubernetes.io/projected/0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7-kube-api-access-6gp72\") pod \"metallb-operator-controller-manager-57b788c8f7-tgv57\" (UID: \"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7\") " pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.213170 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-4d57n" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.213673 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.221440 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.275982 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8"] Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.281975 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-webhook-cert\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.282112 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhpvb\" (UniqueName: \"kubernetes.io/projected/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-kube-api-access-mhpvb\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.282239 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-apiservice-cert\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.287771 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.383908 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-webhook-cert\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.383969 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhpvb\" (UniqueName: \"kubernetes.io/projected/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-kube-api-access-mhpvb\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.384028 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-apiservice-cert\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.387924 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-webhook-cert\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.388579 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-apiservice-cert\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.407464 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhpvb\" (UniqueName: \"kubernetes.io/projected/c5b1e90c-cba0-40f0-b18a-4b7c8c52900c-kube-api-access-mhpvb\") pod \"metallb-operator-webhook-server-86896cd6d4-h6dq8\" (UID: \"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c\") " pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.589135 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.756809 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57"] Dec 04 06:24:30 crc kubenswrapper[4685]: I1204 06:24:30.767894 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:31 crc kubenswrapper[4685]: I1204 06:24:31.022891 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8"] Dec 04 06:24:31 crc kubenswrapper[4685]: W1204 06:24:31.035173 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5b1e90c_cba0_40f0_b18a_4b7c8c52900c.slice/crio-9533772148f616eb6167786ea21e72c29f5f62cb9e9e99e0bcb0969ad5a4725f WatchSource:0}: Error finding container 9533772148f616eb6167786ea21e72c29f5f62cb9e9e99e0bcb0969ad5a4725f: Status 404 returned error can't find the container with id 9533772148f616eb6167786ea21e72c29f5f62cb9e9e99e0bcb0969ad5a4725f Dec 04 06:24:31 crc kubenswrapper[4685]: I1204 06:24:31.734141 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" event={"ID":"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7","Type":"ContainerStarted","Data":"af7b3a70db1c31543a97ff68dba5f6ad4c207c5abe6e66136b3221cb4ffe0353"} Dec 04 06:24:31 crc kubenswrapper[4685]: I1204 06:24:31.735524 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" event={"ID":"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c","Type":"ContainerStarted","Data":"9533772148f616eb6167786ea21e72c29f5f62cb9e9e99e0bcb0969ad5a4725f"} Dec 04 06:24:33 crc kubenswrapper[4685]: I1204 06:24:33.731334 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcfh4"] Dec 04 06:24:33 crc kubenswrapper[4685]: I1204 06:24:33.750849 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qcfh4" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="registry-server" containerID="cri-o://20d25ab279a3efe58a23f67c5d9c2cd459daf78f92b780250b16f04b8c03692c" gracePeriod=2 Dec 04 06:24:34 crc kubenswrapper[4685]: I1204 06:24:34.764582 4685 generic.go:334] "Generic (PLEG): container finished" podID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerID="20d25ab279a3efe58a23f67c5d9c2cd459daf78f92b780250b16f04b8c03692c" exitCode=0 Dec 04 06:24:34 crc kubenswrapper[4685]: I1204 06:24:34.764674 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerDied","Data":"20d25ab279a3efe58a23f67c5d9c2cd459daf78f92b780250b16f04b8c03692c"} Dec 04 06:24:35 crc kubenswrapper[4685]: I1204 06:24:35.870443 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:35 crc kubenswrapper[4685]: I1204 06:24:35.990072 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-catalog-content\") pod \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " Dec 04 06:24:35 crc kubenswrapper[4685]: I1204 06:24:35.990123 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-utilities\") pod \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " Dec 04 06:24:35 crc kubenswrapper[4685]: I1204 06:24:35.990196 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9vq7\" (UniqueName: \"kubernetes.io/projected/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-kube-api-access-h9vq7\") pod \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\" (UID: \"17eb5a3d-e7a6-4b3a-867c-af67fb887a94\") " Dec 04 06:24:35 crc kubenswrapper[4685]: I1204 06:24:35.992510 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-utilities" (OuterVolumeSpecName: "utilities") pod "17eb5a3d-e7a6-4b3a-867c-af67fb887a94" (UID: "17eb5a3d-e7a6-4b3a-867c-af67fb887a94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.009350 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-kube-api-access-h9vq7" (OuterVolumeSpecName: "kube-api-access-h9vq7") pod "17eb5a3d-e7a6-4b3a-867c-af67fb887a94" (UID: "17eb5a3d-e7a6-4b3a-867c-af67fb887a94"). InnerVolumeSpecName "kube-api-access-h9vq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.092156 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.092529 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9vq7\" (UniqueName: \"kubernetes.io/projected/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-kube-api-access-h9vq7\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.096088 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17eb5a3d-e7a6-4b3a-867c-af67fb887a94" (UID: "17eb5a3d-e7a6-4b3a-867c-af67fb887a94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.193837 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17eb5a3d-e7a6-4b3a-867c-af67fb887a94-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.785330 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" event={"ID":"0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7","Type":"ContainerStarted","Data":"bababa1ae83f8319c9473808abf5ccdb6570418939dc1c3fc95883c1a51dffb3"} Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.789054 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcfh4" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.789010 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcfh4" event={"ID":"17eb5a3d-e7a6-4b3a-867c-af67fb887a94","Type":"ContainerDied","Data":"557ae59bf971e12f4d3e5fd5328fca7d8483422c6f0c7f051de1c6c634ee1b98"} Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.789352 4685 scope.go:117] "RemoveContainer" containerID="20d25ab279a3efe58a23f67c5d9c2cd459daf78f92b780250b16f04b8c03692c" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.791336 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" event={"ID":"c5b1e90c-cba0-40f0-b18a-4b7c8c52900c","Type":"ContainerStarted","Data":"bcec94fbf6509f4da2c16c5e80a09db881b9464bad47965b236d24d73d353ae3"} Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.791847 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.820834 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" podStartSLOduration=4.386355874 podStartE2EDuration="7.820802119s" podCreationTimestamp="2025-12-04 06:24:29 +0000 UTC" firstStartedPulling="2025-12-04 06:24:30.766637051 +0000 UTC m=+807.954867826" lastFinishedPulling="2025-12-04 06:24:34.201083296 +0000 UTC m=+811.389314071" observedRunningTime="2025-12-04 06:24:36.819113544 +0000 UTC m=+814.007344349" watchObservedRunningTime="2025-12-04 06:24:36.820802119 +0000 UTC m=+814.009032924" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.826033 4685 scope.go:117] "RemoveContainer" containerID="5419af0a92cb9c6171193d6f6a8cee23a77f4970a1c3f7f7a8ccecdebc264854" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.853368 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" podStartSLOduration=2.014035125 podStartE2EDuration="6.853348998s" podCreationTimestamp="2025-12-04 06:24:30 +0000 UTC" firstStartedPulling="2025-12-04 06:24:31.041628001 +0000 UTC m=+808.229858816" lastFinishedPulling="2025-12-04 06:24:35.880941914 +0000 UTC m=+813.069172689" observedRunningTime="2025-12-04 06:24:36.849624698 +0000 UTC m=+814.037855483" watchObservedRunningTime="2025-12-04 06:24:36.853348998 +0000 UTC m=+814.041579783" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.876776 4685 scope.go:117] "RemoveContainer" containerID="721a67eb14ce81cefa0e807745c1e3eed69cbee70a0c2b36660f57a9ee3a11ae" Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.879084 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcfh4"] Dec 04 06:24:36 crc kubenswrapper[4685]: I1204 06:24:36.883954 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qcfh4"] Dec 04 06:24:37 crc kubenswrapper[4685]: I1204 06:24:37.132862 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" path="/var/lib/kubelet/pods/17eb5a3d-e7a6-4b3a-867c-af67fb887a94/volumes" Dec 04 06:24:37 crc kubenswrapper[4685]: I1204 06:24:37.798002 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:24:50 crc kubenswrapper[4685]: I1204 06:24:50.600216 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-86896cd6d4-h6dq8" Dec 04 06:25:10 crc kubenswrapper[4685]: I1204 06:25:10.292237 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-57b788c8f7-tgv57" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.105285 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv"] Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.106051 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="extract-utilities" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.106079 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="extract-utilities" Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.106096 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="registry-server" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.106107 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="registry-server" Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.106124 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="extract-content" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.106137 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="extract-content" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.106315 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="17eb5a3d-e7a6-4b3a-867c-af67fb887a94" containerName="registry-server" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.106976 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.110568 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.111075 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-nh8gq" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.135201 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bf5dr"] Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.155913 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv"] Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.156517 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.168638 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.169510 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.205111 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vglm9\" (UniqueName: \"kubernetes.io/projected/4eb6708c-7ef6-477d-a404-fd591df256b5-kube-api-access-vglm9\") pod \"frr-k8s-webhook-server-7fcb986d4-hcsrv\" (UID: \"4eb6708c-7ef6-477d-a404-fd591df256b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.205200 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4eb6708c-7ef6-477d-a404-fd591df256b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-hcsrv\" (UID: \"4eb6708c-7ef6-477d-a404-fd591df256b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.209146 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5mm56"] Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.209995 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.215766 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.215948 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.216024 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.219679 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-s454s" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.256961 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-d7jj8"] Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.257835 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.262535 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.277087 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d7jj8"] Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309223 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309268 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metallb-excludel2\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309289 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-reloader\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309304 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-conf\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309328 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vglm9\" (UniqueName: \"kubernetes.io/projected/4eb6708c-7ef6-477d-a404-fd591df256b5-kube-api-access-vglm9\") pod \"frr-k8s-webhook-server-7fcb986d4-hcsrv\" (UID: \"4eb6708c-7ef6-477d-a404-fd591df256b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309352 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-metrics-certs\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309371 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klf82\" (UniqueName: \"kubernetes.io/projected/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-kube-api-access-klf82\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309394 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metrics-certs\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309422 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-startup\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309440 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r526c\" (UniqueName: \"kubernetes.io/projected/7ebc5585-22bd-459e-bd2d-db1bc027a03f-kube-api-access-r526c\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309476 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-metrics\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309493 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4eb6708c-7ef6-477d-a404-fd591df256b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-hcsrv\" (UID: \"4eb6708c-7ef6-477d-a404-fd591df256b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.309512 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-sockets\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.321395 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4eb6708c-7ef6-477d-a404-fd591df256b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-hcsrv\" (UID: \"4eb6708c-7ef6-477d-a404-fd591df256b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.336424 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vglm9\" (UniqueName: \"kubernetes.io/projected/4eb6708c-7ef6-477d-a404-fd591df256b5-kube-api-access-vglm9\") pod \"frr-k8s-webhook-server-7fcb986d4-hcsrv\" (UID: \"4eb6708c-7ef6-477d-a404-fd591df256b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411248 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-metrics\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411309 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf7lx\" (UniqueName: \"kubernetes.io/projected/62ade17b-938c-44f3-9af9-b5138439e87e-kube-api-access-cf7lx\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411335 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-sockets\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411367 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411390 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metallb-excludel2\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411429 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-reloader\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411450 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-conf\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411483 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-metrics-certs\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411520 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klf82\" (UniqueName: \"kubernetes.io/projected/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-kube-api-access-klf82\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411553 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-startup\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metrics-certs\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411602 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62ade17b-938c-44f3-9af9-b5138439e87e-metrics-certs\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411634 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r526c\" (UniqueName: \"kubernetes.io/projected/7ebc5585-22bd-459e-bd2d-db1bc027a03f-kube-api-access-r526c\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.411660 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62ade17b-938c-44f3-9af9-b5138439e87e-cert\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.412167 4685 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.412311 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metrics-certs podName:7ebc5585-22bd-459e-bd2d-db1bc027a03f nodeName:}" failed. No retries permitted until 2025-12-04 06:25:11.912286703 +0000 UTC m=+849.100517478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metrics-certs") pod "speaker-5mm56" (UID: "7ebc5585-22bd-459e-bd2d-db1bc027a03f") : secret "speaker-certs-secret" not found Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.412433 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-conf\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.412539 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-sockets\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.412567 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-metrics\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.412794 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-reloader\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.413237 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metallb-excludel2\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.413335 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-frr-startup\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.412494 4685 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.414561 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist podName:7ebc5585-22bd-459e-bd2d-db1bc027a03f nodeName:}" failed. No retries permitted until 2025-12-04 06:25:11.914543355 +0000 UTC m=+849.102774130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist") pod "speaker-5mm56" (UID: "7ebc5585-22bd-459e-bd2d-db1bc027a03f") : secret "metallb-memberlist" not found Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.415060 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-metrics-certs\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.433848 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klf82\" (UniqueName: \"kubernetes.io/projected/e8ff674f-4d33-4a9f-9dca-c6c7cc63d394-kube-api-access-klf82\") pod \"frr-k8s-bf5dr\" (UID: \"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394\") " pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.434671 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.438223 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r526c\" (UniqueName: \"kubernetes.io/projected/7ebc5585-22bd-459e-bd2d-db1bc027a03f-kube-api-access-r526c\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.496243 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.512972 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62ade17b-938c-44f3-9af9-b5138439e87e-metrics-certs\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.513021 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62ade17b-938c-44f3-9af9-b5138439e87e-cert\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.513074 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf7lx\" (UniqueName: \"kubernetes.io/projected/62ade17b-938c-44f3-9af9-b5138439e87e-kube-api-access-cf7lx\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.516518 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.518240 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62ade17b-938c-44f3-9af9-b5138439e87e-metrics-certs\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.534156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62ade17b-938c-44f3-9af9-b5138439e87e-cert\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.538741 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf7lx\" (UniqueName: \"kubernetes.io/projected/62ade17b-938c-44f3-9af9-b5138439e87e-kube-api-access-cf7lx\") pod \"controller-f8648f98b-d7jj8\" (UID: \"62ade17b-938c-44f3-9af9-b5138439e87e\") " pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.573622 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.658143 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv"] Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.790348 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d7jj8"] Dec 04 06:25:11 crc kubenswrapper[4685]: W1204 06:25:11.795589 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62ade17b_938c_44f3_9af9_b5138439e87e.slice/crio-2c07de3179829c9c9ad4a10a1cfede435257c999a38b4dc3b261136443259755 WatchSource:0}: Error finding container 2c07de3179829c9c9ad4a10a1cfede435257c999a38b4dc3b261136443259755: Status 404 returned error can't find the container with id 2c07de3179829c9c9ad4a10a1cfede435257c999a38b4dc3b261136443259755 Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.918512 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metrics-certs\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.919221 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.919486 4685 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 06:25:11 crc kubenswrapper[4685]: E1204 06:25:11.919595 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist podName:7ebc5585-22bd-459e-bd2d-db1bc027a03f nodeName:}" failed. No retries permitted until 2025-12-04 06:25:12.919560566 +0000 UTC m=+850.107791381 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist") pod "speaker-5mm56" (UID: "7ebc5585-22bd-459e-bd2d-db1bc027a03f") : secret "metallb-memberlist" not found Dec 04 06:25:11 crc kubenswrapper[4685]: I1204 06:25:11.925691 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-metrics-certs\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:12 crc kubenswrapper[4685]: I1204 06:25:12.026047 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"6f46d857a988b40f422f6d709eb25ffa9e76c02fd22591d89a3198c0aff9ddb2"} Dec 04 06:25:12 crc kubenswrapper[4685]: I1204 06:25:12.027803 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d7jj8" event={"ID":"62ade17b-938c-44f3-9af9-b5138439e87e","Type":"ContainerStarted","Data":"c36f6108d77529589e1bb3f03d55f8afc2390aeb5869569e5320ce09ad4dc234"} Dec 04 06:25:12 crc kubenswrapper[4685]: I1204 06:25:12.027877 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d7jj8" event={"ID":"62ade17b-938c-44f3-9af9-b5138439e87e","Type":"ContainerStarted","Data":"2c07de3179829c9c9ad4a10a1cfede435257c999a38b4dc3b261136443259755"} Dec 04 06:25:12 crc kubenswrapper[4685]: I1204 06:25:12.028918 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" event={"ID":"4eb6708c-7ef6-477d-a404-fd591df256b5","Type":"ContainerStarted","Data":"ee8b6f29127312bce0355b3db255358774525ba90deaa44d26541095a4122f24"} Dec 04 06:25:12 crc kubenswrapper[4685]: I1204 06:25:12.935541 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:12 crc kubenswrapper[4685]: I1204 06:25:12.947594 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7ebc5585-22bd-459e-bd2d-db1bc027a03f-memberlist\") pod \"speaker-5mm56\" (UID: \"7ebc5585-22bd-459e-bd2d-db1bc027a03f\") " pod="metallb-system/speaker-5mm56" Dec 04 06:25:13 crc kubenswrapper[4685]: I1204 06:25:13.037997 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5mm56" Dec 04 06:25:13 crc kubenswrapper[4685]: I1204 06:25:13.046317 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d7jj8" event={"ID":"62ade17b-938c-44f3-9af9-b5138439e87e","Type":"ContainerStarted","Data":"282a59a9c6bfd719867b61f5c8a98475a590640e1748e1775d9165b55b1368a3"} Dec 04 06:25:13 crc kubenswrapper[4685]: I1204 06:25:13.047224 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:13 crc kubenswrapper[4685]: I1204 06:25:13.070830 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-d7jj8" podStartSLOduration=2.07080581 podStartE2EDuration="2.07080581s" podCreationTimestamp="2025-12-04 06:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:25:13.067708611 +0000 UTC m=+850.255939406" watchObservedRunningTime="2025-12-04 06:25:13.07080581 +0000 UTC m=+850.259036585" Dec 04 06:25:14 crc kubenswrapper[4685]: I1204 06:25:14.080730 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5mm56" event={"ID":"7ebc5585-22bd-459e-bd2d-db1bc027a03f","Type":"ContainerStarted","Data":"a9d2ad34946e34c0054c53f08793780a378ed472e65501dccde5e4a1d431dd7a"} Dec 04 06:25:14 crc kubenswrapper[4685]: I1204 06:25:14.081168 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5mm56" event={"ID":"7ebc5585-22bd-459e-bd2d-db1bc027a03f","Type":"ContainerStarted","Data":"ead47ee1a2da8c116b6e3941b069ef790fa2cdac7f02b2180675f87aa14355f6"} Dec 04 06:25:14 crc kubenswrapper[4685]: I1204 06:25:14.081185 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5mm56" event={"ID":"7ebc5585-22bd-459e-bd2d-db1bc027a03f","Type":"ContainerStarted","Data":"0b26497acab87e0a3e33832acb53008f93fe24455dc52c14972f426e8c450683"} Dec 04 06:25:14 crc kubenswrapper[4685]: I1204 06:25:14.081582 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5mm56" Dec 04 06:25:14 crc kubenswrapper[4685]: I1204 06:25:14.107738 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5mm56" podStartSLOduration=3.107714927 podStartE2EDuration="3.107714927s" podCreationTimestamp="2025-12-04 06:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:25:14.103708309 +0000 UTC m=+851.291939104" watchObservedRunningTime="2025-12-04 06:25:14.107714927 +0000 UTC m=+851.295945702" Dec 04 06:25:20 crc kubenswrapper[4685]: I1204 06:25:20.125201 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" event={"ID":"4eb6708c-7ef6-477d-a404-fd591df256b5","Type":"ContainerStarted","Data":"b069005bb379a2b33ad9390352faab983cf067bc686beea5465e11aec660977c"} Dec 04 06:25:20 crc kubenswrapper[4685]: I1204 06:25:20.125912 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:20 crc kubenswrapper[4685]: I1204 06:25:20.127555 4685 generic.go:334] "Generic (PLEG): container finished" podID="e8ff674f-4d33-4a9f-9dca-c6c7cc63d394" containerID="018f22f25ce5ff68dfb5e9040cb7c3a13d1ba9512f330f9c9b07c18d7ee68e96" exitCode=0 Dec 04 06:25:20 crc kubenswrapper[4685]: I1204 06:25:20.127584 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerDied","Data":"018f22f25ce5ff68dfb5e9040cb7c3a13d1ba9512f330f9c9b07c18d7ee68e96"} Dec 04 06:25:20 crc kubenswrapper[4685]: I1204 06:25:20.146689 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" podStartSLOduration=1.257351185 podStartE2EDuration="9.14666723s" podCreationTimestamp="2025-12-04 06:25:11 +0000 UTC" firstStartedPulling="2025-12-04 06:25:11.665447379 +0000 UTC m=+848.853678164" lastFinishedPulling="2025-12-04 06:25:19.554763394 +0000 UTC m=+856.742994209" observedRunningTime="2025-12-04 06:25:20.142175167 +0000 UTC m=+857.330405932" watchObservedRunningTime="2025-12-04 06:25:20.14666723 +0000 UTC m=+857.334898015" Dec 04 06:25:21 crc kubenswrapper[4685]: I1204 06:25:21.139762 4685 generic.go:334] "Generic (PLEG): container finished" podID="e8ff674f-4d33-4a9f-9dca-c6c7cc63d394" containerID="b212a59eb0d5b64d486c68e3a4c39dc182a3386f31ba62a775f744295a9c42d7" exitCode=0 Dec 04 06:25:21 crc kubenswrapper[4685]: I1204 06:25:21.139859 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerDied","Data":"b212a59eb0d5b64d486c68e3a4c39dc182a3386f31ba62a775f744295a9c42d7"} Dec 04 06:25:21 crc kubenswrapper[4685]: I1204 06:25:21.579384 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-d7jj8" Dec 04 06:25:22 crc kubenswrapper[4685]: I1204 06:25:22.150284 4685 generic.go:334] "Generic (PLEG): container finished" podID="e8ff674f-4d33-4a9f-9dca-c6c7cc63d394" containerID="98792052d5060d14af3a24a86a5e5863ce85dc5f58243cdf7fdef7b130babc44" exitCode=0 Dec 04 06:25:22 crc kubenswrapper[4685]: I1204 06:25:22.150339 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerDied","Data":"98792052d5060d14af3a24a86a5e5863ce85dc5f58243cdf7fdef7b130babc44"} Dec 04 06:25:23 crc kubenswrapper[4685]: I1204 06:25:23.042959 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5mm56" Dec 04 06:25:23 crc kubenswrapper[4685]: I1204 06:25:23.159799 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"ba0265be529101183a348e363a921b9a82277140be97cdf44e8bc18076eac8e9"} Dec 04 06:25:23 crc kubenswrapper[4685]: I1204 06:25:23.159858 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"0404960ee184b859b92fc82f1a085fa2f9dfb1e558c777f5adbb291c3a7a817e"} Dec 04 06:25:23 crc kubenswrapper[4685]: I1204 06:25:23.159869 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"5a9f1283d2347c6b92f4ba0a0cd0c6f5fbcb7b5a8c41d7a26c2db68689288cce"} Dec 04 06:25:23 crc kubenswrapper[4685]: I1204 06:25:23.159877 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"4cb5da064cd76fdc9f8253ea0ff970adcea9dc20017abf27373f7dadc6a95275"} Dec 04 06:25:23 crc kubenswrapper[4685]: I1204 06:25:23.159889 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"a1d777525f063e8a7ea37aeedf67a8c0dab1115d2b861b7ece1273255ed91647"} Dec 04 06:25:24 crc kubenswrapper[4685]: I1204 06:25:24.173925 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bf5dr" event={"ID":"e8ff674f-4d33-4a9f-9dca-c6c7cc63d394","Type":"ContainerStarted","Data":"5901391d9b2d68f6f527d0f30ce5620da0eccfc794de7bca9f28e2ad578c2309"} Dec 04 06:25:24 crc kubenswrapper[4685]: I1204 06:25:24.174591 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:24 crc kubenswrapper[4685]: I1204 06:25:24.204557 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bf5dr" podStartSLOduration=5.333804226 podStartE2EDuration="13.204526609s" podCreationTimestamp="2025-12-04 06:25:11 +0000 UTC" firstStartedPulling="2025-12-04 06:25:11.645228616 +0000 UTC m=+848.833459391" lastFinishedPulling="2025-12-04 06:25:19.515950999 +0000 UTC m=+856.704181774" observedRunningTime="2025-12-04 06:25:24.201460213 +0000 UTC m=+861.389690988" watchObservedRunningTime="2025-12-04 06:25:24.204526609 +0000 UTC m=+861.392757394" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.142599 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qd8g8"] Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.143994 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.174919 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd8g8"] Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.304931 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55sjl\" (UniqueName: \"kubernetes.io/projected/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-kube-api-access-55sjl\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.305260 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-catalog-content\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.305355 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-utilities\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.406858 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-catalog-content\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.406931 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-utilities\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.407008 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55sjl\" (UniqueName: \"kubernetes.io/projected/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-kube-api-access-55sjl\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.407490 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-catalog-content\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.407741 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-utilities\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.436208 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55sjl\" (UniqueName: \"kubernetes.io/projected/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-kube-api-access-55sjl\") pod \"community-operators-qd8g8\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.463174 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:25 crc kubenswrapper[4685]: I1204 06:25:25.955528 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd8g8"] Dec 04 06:25:26 crc kubenswrapper[4685]: I1204 06:25:26.188202 4685 generic.go:334] "Generic (PLEG): container finished" podID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerID="f3f883d2b5e7a9b2f7df8198972fbf818acf4edca883bc5a376c25b5035e3737" exitCode=0 Dec 04 06:25:26 crc kubenswrapper[4685]: I1204 06:25:26.188255 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerDied","Data":"f3f883d2b5e7a9b2f7df8198972fbf818acf4edca883bc5a376c25b5035e3737"} Dec 04 06:25:26 crc kubenswrapper[4685]: I1204 06:25:26.188283 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerStarted","Data":"950bf371eb315d86713e104658df00cd1c3f9bfb02415d280d6950f90942cde8"} Dec 04 06:25:26 crc kubenswrapper[4685]: I1204 06:25:26.496675 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:26 crc kubenswrapper[4685]: I1204 06:25:26.538661 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:27 crc kubenswrapper[4685]: I1204 06:25:27.197357 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerStarted","Data":"1bfecf2778d2c84b330e560126d7680500fa6d1dcb80260f64e9ee47b8c49fc7"} Dec 04 06:25:28 crc kubenswrapper[4685]: I1204 06:25:28.205473 4685 generic.go:334] "Generic (PLEG): container finished" podID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerID="1bfecf2778d2c84b330e560126d7680500fa6d1dcb80260f64e9ee47b8c49fc7" exitCode=0 Dec 04 06:25:28 crc kubenswrapper[4685]: I1204 06:25:28.205532 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerDied","Data":"1bfecf2778d2c84b330e560126d7680500fa6d1dcb80260f64e9ee47b8c49fc7"} Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.217625 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerStarted","Data":"441e8ffebfa9ac4d186e39bfdf5018b098014dce1b2fb266b26ab2b54be74f41"} Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.244152 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qd8g8" podStartSLOduration=1.79928526 podStartE2EDuration="4.244117s" podCreationTimestamp="2025-12-04 06:25:25 +0000 UTC" firstStartedPulling="2025-12-04 06:25:26.189655901 +0000 UTC m=+863.377886676" lastFinishedPulling="2025-12-04 06:25:28.634487631 +0000 UTC m=+865.822718416" observedRunningTime="2025-12-04 06:25:29.239800623 +0000 UTC m=+866.428031418" watchObservedRunningTime="2025-12-04 06:25:29.244117 +0000 UTC m=+866.432347775" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.500096 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kggwz"] Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.501277 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.509357 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-tb6dp" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.510277 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.515209 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.519764 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kggwz"] Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.679827 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fg8b\" (UniqueName: \"kubernetes.io/projected/34e918f6-85f8-4b9b-b135-35b0d28bab97-kube-api-access-6fg8b\") pod \"openstack-operator-index-kggwz\" (UID: \"34e918f6-85f8-4b9b-b135-35b0d28bab97\") " pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.781226 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fg8b\" (UniqueName: \"kubernetes.io/projected/34e918f6-85f8-4b9b-b135-35b0d28bab97-kube-api-access-6fg8b\") pod \"openstack-operator-index-kggwz\" (UID: \"34e918f6-85f8-4b9b-b135-35b0d28bab97\") " pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.811024 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fg8b\" (UniqueName: \"kubernetes.io/projected/34e918f6-85f8-4b9b-b135-35b0d28bab97-kube-api-access-6fg8b\") pod \"openstack-operator-index-kggwz\" (UID: \"34e918f6-85f8-4b9b-b135-35b0d28bab97\") " pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:29 crc kubenswrapper[4685]: I1204 06:25:29.816842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:30 crc kubenswrapper[4685]: I1204 06:25:30.255655 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kggwz"] Dec 04 06:25:30 crc kubenswrapper[4685]: W1204 06:25:30.263172 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34e918f6_85f8_4b9b_b135_35b0d28bab97.slice/crio-f9d52310a3086e0fa93fbf8963c5ee9e280b596194c80e0f74f75c83c50d6727 WatchSource:0}: Error finding container f9d52310a3086e0fa93fbf8963c5ee9e280b596194c80e0f74f75c83c50d6727: Status 404 returned error can't find the container with id f9d52310a3086e0fa93fbf8963c5ee9e280b596194c80e0f74f75c83c50d6727 Dec 04 06:25:31 crc kubenswrapper[4685]: I1204 06:25:31.240793 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kggwz" event={"ID":"34e918f6-85f8-4b9b-b135-35b0d28bab97","Type":"ContainerStarted","Data":"f9d52310a3086e0fa93fbf8963c5ee9e280b596194c80e0f74f75c83c50d6727"} Dec 04 06:25:31 crc kubenswrapper[4685]: I1204 06:25:31.444359 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hcsrv" Dec 04 06:25:33 crc kubenswrapper[4685]: I1204 06:25:33.272050 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kggwz" event={"ID":"34e918f6-85f8-4b9b-b135-35b0d28bab97","Type":"ContainerStarted","Data":"d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd"} Dec 04 06:25:33 crc kubenswrapper[4685]: I1204 06:25:33.296067 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kggwz" podStartSLOduration=1.5283337270000001 podStartE2EDuration="4.296044032s" podCreationTimestamp="2025-12-04 06:25:29 +0000 UTC" firstStartedPulling="2025-12-04 06:25:30.265522704 +0000 UTC m=+867.453753509" lastFinishedPulling="2025-12-04 06:25:33.033233029 +0000 UTC m=+870.221463814" observedRunningTime="2025-12-04 06:25:33.286023933 +0000 UTC m=+870.474254708" watchObservedRunningTime="2025-12-04 06:25:33.296044032 +0000 UTC m=+870.484274827" Dec 04 06:25:34 crc kubenswrapper[4685]: I1204 06:25:34.694354 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kggwz"] Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.099169 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7m62c"] Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.100161 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.116202 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7m62c"] Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.173919 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl6tm\" (UniqueName: \"kubernetes.io/projected/569ea63b-9c26-411f-a899-6b881743135f-kube-api-access-wl6tm\") pod \"openstack-operator-index-7m62c\" (UID: \"569ea63b-9c26-411f-a899-6b881743135f\") " pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.275959 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl6tm\" (UniqueName: \"kubernetes.io/projected/569ea63b-9c26-411f-a899-6b881743135f-kube-api-access-wl6tm\") pod \"openstack-operator-index-7m62c\" (UID: \"569ea63b-9c26-411f-a899-6b881743135f\") " pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.281539 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-kggwz" podUID="34e918f6-85f8-4b9b-b135-35b0d28bab97" containerName="registry-server" containerID="cri-o://d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd" gracePeriod=2 Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.297519 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl6tm\" (UniqueName: \"kubernetes.io/projected/569ea63b-9c26-411f-a899-6b881743135f-kube-api-access-wl6tm\") pod \"openstack-operator-index-7m62c\" (UID: \"569ea63b-9c26-411f-a899-6b881743135f\") " pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.458111 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.463340 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.463395 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.545571 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.711791 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.722697 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7m62c"] Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.786994 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fg8b\" (UniqueName: \"kubernetes.io/projected/34e918f6-85f8-4b9b-b135-35b0d28bab97-kube-api-access-6fg8b\") pod \"34e918f6-85f8-4b9b-b135-35b0d28bab97\" (UID: \"34e918f6-85f8-4b9b-b135-35b0d28bab97\") " Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.791393 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e918f6-85f8-4b9b-b135-35b0d28bab97-kube-api-access-6fg8b" (OuterVolumeSpecName: "kube-api-access-6fg8b") pod "34e918f6-85f8-4b9b-b135-35b0d28bab97" (UID: "34e918f6-85f8-4b9b-b135-35b0d28bab97"). InnerVolumeSpecName "kube-api-access-6fg8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:25:35 crc kubenswrapper[4685]: I1204 06:25:35.889007 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fg8b\" (UniqueName: \"kubernetes.io/projected/34e918f6-85f8-4b9b-b135-35b0d28bab97-kube-api-access-6fg8b\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.290292 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7m62c" event={"ID":"569ea63b-9c26-411f-a899-6b881743135f","Type":"ContainerStarted","Data":"8b0847627fbe58886da07bc7d7cb8571aca22cbf057c53c42045446563a00550"} Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.290732 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7m62c" event={"ID":"569ea63b-9c26-411f-a899-6b881743135f","Type":"ContainerStarted","Data":"ebf3ac51bdb3a68634ea2307eb8e027ab0251a429dfb0867610790ceebdb198f"} Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.292231 4685 generic.go:334] "Generic (PLEG): container finished" podID="34e918f6-85f8-4b9b-b135-35b0d28bab97" containerID="d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd" exitCode=0 Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.292283 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kggwz" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.292312 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kggwz" event={"ID":"34e918f6-85f8-4b9b-b135-35b0d28bab97","Type":"ContainerDied","Data":"d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd"} Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.292393 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kggwz" event={"ID":"34e918f6-85f8-4b9b-b135-35b0d28bab97","Type":"ContainerDied","Data":"f9d52310a3086e0fa93fbf8963c5ee9e280b596194c80e0f74f75c83c50d6727"} Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.292439 4685 scope.go:117] "RemoveContainer" containerID="d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.310649 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7m62c" podStartSLOduration=1.258550264 podStartE2EDuration="1.310627372s" podCreationTimestamp="2025-12-04 06:25:35 +0000 UTC" firstStartedPulling="2025-12-04 06:25:35.722895359 +0000 UTC m=+872.911126134" lastFinishedPulling="2025-12-04 06:25:35.774972467 +0000 UTC m=+872.963203242" observedRunningTime="2025-12-04 06:25:36.309151736 +0000 UTC m=+873.497382521" watchObservedRunningTime="2025-12-04 06:25:36.310627372 +0000 UTC m=+873.498858187" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.354638 4685 scope.go:117] "RemoveContainer" containerID="d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.354936 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:36 crc kubenswrapper[4685]: E1204 06:25:36.357323 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd\": container with ID starting with d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd not found: ID does not exist" containerID="d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.357379 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd"} err="failed to get container status \"d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd\": rpc error: code = NotFound desc = could not find container \"d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd\": container with ID starting with d4e8a8fd7cee2ddc6cb56fd13fe967a0c16c137c6f8b20c7632e931b8d72d3dd not found: ID does not exist" Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.362626 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kggwz"] Dec 04 06:25:36 crc kubenswrapper[4685]: I1204 06:25:36.365719 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-kggwz"] Dec 04 06:25:37 crc kubenswrapper[4685]: I1204 06:25:37.142304 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34e918f6-85f8-4b9b-b135-35b0d28bab97" path="/var/lib/kubelet/pods/34e918f6-85f8-4b9b-b135-35b0d28bab97/volumes" Dec 04 06:25:39 crc kubenswrapper[4685]: I1204 06:25:39.085861 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qd8g8"] Dec 04 06:25:39 crc kubenswrapper[4685]: I1204 06:25:39.086531 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qd8g8" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="registry-server" containerID="cri-o://441e8ffebfa9ac4d186e39bfdf5018b098014dce1b2fb266b26ab2b54be74f41" gracePeriod=2 Dec 04 06:25:39 crc kubenswrapper[4685]: I1204 06:25:39.319692 4685 generic.go:334] "Generic (PLEG): container finished" podID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerID="441e8ffebfa9ac4d186e39bfdf5018b098014dce1b2fb266b26ab2b54be74f41" exitCode=0 Dec 04 06:25:39 crc kubenswrapper[4685]: I1204 06:25:39.319780 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerDied","Data":"441e8ffebfa9ac4d186e39bfdf5018b098014dce1b2fb266b26ab2b54be74f41"} Dec 04 06:25:41 crc kubenswrapper[4685]: I1204 06:25:41.499220 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bf5dr" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:41.991145 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.053295 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55sjl\" (UniqueName: \"kubernetes.io/projected/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-kube-api-access-55sjl\") pod \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.053434 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-catalog-content\") pod \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.053483 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-utilities\") pod \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\" (UID: \"2c4a96ce-268f-42ca-a7f0-f50aafd58f89\") " Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.054471 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-utilities" (OuterVolumeSpecName: "utilities") pod "2c4a96ce-268f-42ca-a7f0-f50aafd58f89" (UID: "2c4a96ce-268f-42ca-a7f0-f50aafd58f89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.059303 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-kube-api-access-55sjl" (OuterVolumeSpecName: "kube-api-access-55sjl") pod "2c4a96ce-268f-42ca-a7f0-f50aafd58f89" (UID: "2c4a96ce-268f-42ca-a7f0-f50aafd58f89"). InnerVolumeSpecName "kube-api-access-55sjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.099522 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c4a96ce-268f-42ca-a7f0-f50aafd58f89" (UID: "2c4a96ce-268f-42ca-a7f0-f50aafd58f89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.155086 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.155123 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55sjl\" (UniqueName: \"kubernetes.io/projected/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-kube-api-access-55sjl\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.155133 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4a96ce-268f-42ca-a7f0-f50aafd58f89-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.260828 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd8g8" event={"ID":"2c4a96ce-268f-42ca-a7f0-f50aafd58f89","Type":"ContainerDied","Data":"950bf371eb315d86713e104658df00cd1c3f9bfb02415d280d6950f90942cde8"} Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.260903 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd8g8" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.260908 4685 scope.go:117] "RemoveContainer" containerID="441e8ffebfa9ac4d186e39bfdf5018b098014dce1b2fb266b26ab2b54be74f41" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.299166 4685 scope.go:117] "RemoveContainer" containerID="1bfecf2778d2c84b330e560126d7680500fa6d1dcb80260f64e9ee47b8c49fc7" Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.299997 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qd8g8"] Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.304261 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qd8g8"] Dec 04 06:25:42 crc kubenswrapper[4685]: I1204 06:25:42.332993 4685 scope.go:117] "RemoveContainer" containerID="f3f883d2b5e7a9b2f7df8198972fbf818acf4edca883bc5a376c25b5035e3737" Dec 04 06:25:43 crc kubenswrapper[4685]: I1204 06:25:43.135153 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" path="/var/lib/kubelet/pods/2c4a96ce-268f-42ca-a7f0-f50aafd58f89/volumes" Dec 04 06:25:45 crc kubenswrapper[4685]: I1204 06:25:45.459170 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:45 crc kubenswrapper[4685]: I1204 06:25:45.459221 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:45 crc kubenswrapper[4685]: I1204 06:25:45.505687 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.336263 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-7m62c" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.700612 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f58c7"] Dec 04 06:25:46 crc kubenswrapper[4685]: E1204 06:25:46.701024 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="registry-server" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.701051 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="registry-server" Dec 04 06:25:46 crc kubenswrapper[4685]: E1204 06:25:46.701084 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="extract-utilities" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.701103 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="extract-utilities" Dec 04 06:25:46 crc kubenswrapper[4685]: E1204 06:25:46.701127 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e918f6-85f8-4b9b-b135-35b0d28bab97" containerName="registry-server" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.701142 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e918f6-85f8-4b9b-b135-35b0d28bab97" containerName="registry-server" Dec 04 06:25:46 crc kubenswrapper[4685]: E1204 06:25:46.701168 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="extract-content" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.701183 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="extract-content" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.701482 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e918f6-85f8-4b9b-b135-35b0d28bab97" containerName="registry-server" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.701533 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c4a96ce-268f-42ca-a7f0-f50aafd58f89" containerName="registry-server" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.703397 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.720689 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f58c7"] Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.735586 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-catalog-content\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.735657 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-utilities\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.736182 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l78df\" (UniqueName: \"kubernetes.io/projected/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-kube-api-access-l78df\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.837466 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-catalog-content\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.837547 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-utilities\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.837614 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l78df\" (UniqueName: \"kubernetes.io/projected/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-kube-api-access-l78df\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.838029 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-catalog-content\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.838333 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-utilities\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:46 crc kubenswrapper[4685]: I1204 06:25:46.857255 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l78df\" (UniqueName: \"kubernetes.io/projected/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-kube-api-access-l78df\") pod \"certified-operators-f58c7\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.029502 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.562509 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f58c7"] Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.801427 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv"] Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.810609 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.818192 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9pp5m" Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.820202 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv"] Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.954360 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-bundle\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.954445 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-util\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:47 crc kubenswrapper[4685]: I1204 06:25:47.954592 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59mjh\" (UniqueName: \"kubernetes.io/projected/b19eb475-d7de-4727-8a07-ee8e150204ef-kube-api-access-59mjh\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.056572 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-bundle\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.056641 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-util\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.056715 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59mjh\" (UniqueName: \"kubernetes.io/projected/b19eb475-d7de-4727-8a07-ee8e150204ef-kube-api-access-59mjh\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.057345 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-bundle\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.057345 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-util\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.082624 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59mjh\" (UniqueName: \"kubernetes.io/projected/b19eb475-d7de-4727-8a07-ee8e150204ef-kube-api-access-59mjh\") pod \"a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.135499 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.308914 4685 generic.go:334] "Generic (PLEG): container finished" podID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerID="c39cc1f94aca01975d2321d68386733057d8442a64e9255c858794ab746ed5d6" exitCode=0 Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.309066 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f58c7" event={"ID":"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f","Type":"ContainerDied","Data":"c39cc1f94aca01975d2321d68386733057d8442a64e9255c858794ab746ed5d6"} Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.309353 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f58c7" event={"ID":"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f","Type":"ContainerStarted","Data":"6fbd2b9b8b93b39c238e553506726522519fa1498016aa9e36733352d4619d7a"} Dec 04 06:25:48 crc kubenswrapper[4685]: I1204 06:25:48.648207 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv"] Dec 04 06:25:48 crc kubenswrapper[4685]: W1204 06:25:48.654870 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb19eb475_d7de_4727_8a07_ee8e150204ef.slice/crio-720dc6c92bb1e68d13aaa63741f6cad042396b21c508f94a01ccb96467df90db WatchSource:0}: Error finding container 720dc6c92bb1e68d13aaa63741f6cad042396b21c508f94a01ccb96467df90db: Status 404 returned error can't find the container with id 720dc6c92bb1e68d13aaa63741f6cad042396b21c508f94a01ccb96467df90db Dec 04 06:25:49 crc kubenswrapper[4685]: I1204 06:25:49.317262 4685 generic.go:334] "Generic (PLEG): container finished" podID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerID="c8682bdea2c6435e2a25f59978975ac8d1d07688a765bb5f17cd9e801517b8ff" exitCode=0 Dec 04 06:25:49 crc kubenswrapper[4685]: I1204 06:25:49.317317 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" event={"ID":"b19eb475-d7de-4727-8a07-ee8e150204ef","Type":"ContainerDied","Data":"c8682bdea2c6435e2a25f59978975ac8d1d07688a765bb5f17cd9e801517b8ff"} Dec 04 06:25:49 crc kubenswrapper[4685]: I1204 06:25:49.317628 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" event={"ID":"b19eb475-d7de-4727-8a07-ee8e150204ef","Type":"ContainerStarted","Data":"720dc6c92bb1e68d13aaa63741f6cad042396b21c508f94a01ccb96467df90db"} Dec 04 06:25:49 crc kubenswrapper[4685]: I1204 06:25:49.320526 4685 generic.go:334] "Generic (PLEG): container finished" podID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerID="8a9ac670a1242d05faa3655161c1fdc538ecbf7b68df30cf416994dfeaa07bee" exitCode=0 Dec 04 06:25:49 crc kubenswrapper[4685]: I1204 06:25:49.320554 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f58c7" event={"ID":"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f","Type":"ContainerDied","Data":"8a9ac670a1242d05faa3655161c1fdc538ecbf7b68df30cf416994dfeaa07bee"} Dec 04 06:25:50 crc kubenswrapper[4685]: I1204 06:25:50.338690 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f58c7" event={"ID":"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f","Type":"ContainerStarted","Data":"23290284a93ea74caa6b6721ba921f3609874cca77717c1991824e02ce44378a"} Dec 04 06:25:50 crc kubenswrapper[4685]: I1204 06:25:50.346768 4685 generic.go:334] "Generic (PLEG): container finished" podID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerID="4776051365c46888c624fe0718ae73a24dc2a0dce2f5cd7035a353762243db9d" exitCode=0 Dec 04 06:25:50 crc kubenswrapper[4685]: I1204 06:25:50.346809 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" event={"ID":"b19eb475-d7de-4727-8a07-ee8e150204ef","Type":"ContainerDied","Data":"4776051365c46888c624fe0718ae73a24dc2a0dce2f5cd7035a353762243db9d"} Dec 04 06:25:50 crc kubenswrapper[4685]: I1204 06:25:50.377581 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f58c7" podStartSLOduration=2.843249238 podStartE2EDuration="4.377562913s" podCreationTimestamp="2025-12-04 06:25:46 +0000 UTC" firstStartedPulling="2025-12-04 06:25:48.31161602 +0000 UTC m=+885.499846795" lastFinishedPulling="2025-12-04 06:25:49.845929665 +0000 UTC m=+887.034160470" observedRunningTime="2025-12-04 06:25:50.375097474 +0000 UTC m=+887.563328249" watchObservedRunningTime="2025-12-04 06:25:50.377562913 +0000 UTC m=+887.565793698" Dec 04 06:25:51 crc kubenswrapper[4685]: I1204 06:25:51.354564 4685 generic.go:334] "Generic (PLEG): container finished" podID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerID="8f2b278024cb287b22da42efd14be1c96791dbb1cf88bbca2a4d069baeea2a09" exitCode=0 Dec 04 06:25:51 crc kubenswrapper[4685]: I1204 06:25:51.354848 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" event={"ID":"b19eb475-d7de-4727-8a07-ee8e150204ef","Type":"ContainerDied","Data":"8f2b278024cb287b22da42efd14be1c96791dbb1cf88bbca2a4d069baeea2a09"} Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.713561 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.735197 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-bundle\") pod \"b19eb475-d7de-4727-8a07-ee8e150204ef\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.735299 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-util\") pod \"b19eb475-d7de-4727-8a07-ee8e150204ef\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.735336 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59mjh\" (UniqueName: \"kubernetes.io/projected/b19eb475-d7de-4727-8a07-ee8e150204ef-kube-api-access-59mjh\") pod \"b19eb475-d7de-4727-8a07-ee8e150204ef\" (UID: \"b19eb475-d7de-4727-8a07-ee8e150204ef\") " Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.737510 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-bundle" (OuterVolumeSpecName: "bundle") pod "b19eb475-d7de-4727-8a07-ee8e150204ef" (UID: "b19eb475-d7de-4727-8a07-ee8e150204ef"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.752064 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b19eb475-d7de-4727-8a07-ee8e150204ef-kube-api-access-59mjh" (OuterVolumeSpecName: "kube-api-access-59mjh") pod "b19eb475-d7de-4727-8a07-ee8e150204ef" (UID: "b19eb475-d7de-4727-8a07-ee8e150204ef"). InnerVolumeSpecName "kube-api-access-59mjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.755629 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-util" (OuterVolumeSpecName: "util") pod "b19eb475-d7de-4727-8a07-ee8e150204ef" (UID: "b19eb475-d7de-4727-8a07-ee8e150204ef"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.837832 4685 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.837897 4685 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b19eb475-d7de-4727-8a07-ee8e150204ef-util\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:52 crc kubenswrapper[4685]: I1204 06:25:52.837926 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59mjh\" (UniqueName: \"kubernetes.io/projected/b19eb475-d7de-4727-8a07-ee8e150204ef-kube-api-access-59mjh\") on node \"crc\" DevicePath \"\"" Dec 04 06:25:53 crc kubenswrapper[4685]: I1204 06:25:53.375464 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" event={"ID":"b19eb475-d7de-4727-8a07-ee8e150204ef","Type":"ContainerDied","Data":"720dc6c92bb1e68d13aaa63741f6cad042396b21c508f94a01ccb96467df90db"} Dec 04 06:25:53 crc kubenswrapper[4685]: I1204 06:25:53.375970 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="720dc6c92bb1e68d13aaa63741f6cad042396b21c508f94a01ccb96467df90db" Dec 04 06:25:53 crc kubenswrapper[4685]: I1204 06:25:53.375608 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.360534 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch"] Dec 04 06:25:56 crc kubenswrapper[4685]: E1204 06:25:56.361197 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="util" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.361215 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="util" Dec 04 06:25:56 crc kubenswrapper[4685]: E1204 06:25:56.361227 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="extract" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.361234 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="extract" Dec 04 06:25:56 crc kubenswrapper[4685]: E1204 06:25:56.361250 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="pull" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.361256 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="pull" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.361372 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b19eb475-d7de-4727-8a07-ee8e150204ef" containerName="extract" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.362008 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.365606 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-x2ggc" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.397011 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzsqt\" (UniqueName: \"kubernetes.io/projected/90f5cdd3-8605-4cb5-9aea-83d57954bc17-kube-api-access-wzsqt\") pod \"openstack-operator-controller-operator-778b7fc6d-66vch\" (UID: \"90f5cdd3-8605-4cb5-9aea-83d57954bc17\") " pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.398670 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch"] Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.498503 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzsqt\" (UniqueName: \"kubernetes.io/projected/90f5cdd3-8605-4cb5-9aea-83d57954bc17-kube-api-access-wzsqt\") pod \"openstack-operator-controller-operator-778b7fc6d-66vch\" (UID: \"90f5cdd3-8605-4cb5-9aea-83d57954bc17\") " pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.521315 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzsqt\" (UniqueName: \"kubernetes.io/projected/90f5cdd3-8605-4cb5-9aea-83d57954bc17-kube-api-access-wzsqt\") pod \"openstack-operator-controller-operator-778b7fc6d-66vch\" (UID: \"90f5cdd3-8605-4cb5-9aea-83d57954bc17\") " pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.701437 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:25:56 crc kubenswrapper[4685]: I1204 06:25:56.972645 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch"] Dec 04 06:25:57 crc kubenswrapper[4685]: I1204 06:25:57.030620 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:57 crc kubenswrapper[4685]: I1204 06:25:57.030817 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:57 crc kubenswrapper[4685]: I1204 06:25:57.078695 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:57 crc kubenswrapper[4685]: I1204 06:25:57.412966 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" event={"ID":"90f5cdd3-8605-4cb5-9aea-83d57954bc17","Type":"ContainerStarted","Data":"4973dfc9066821cf696360eff3bf2981be0cb2fcfa2f8e608461923fefbfc034"} Dec 04 06:25:57 crc kubenswrapper[4685]: I1204 06:25:57.468835 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:25:59 crc kubenswrapper[4685]: I1204 06:25:59.893275 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f58c7"] Dec 04 06:26:00 crc kubenswrapper[4685]: I1204 06:26:00.440959 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f58c7" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="registry-server" containerID="cri-o://23290284a93ea74caa6b6721ba921f3609874cca77717c1991824e02ce44378a" gracePeriod=2 Dec 04 06:26:01 crc kubenswrapper[4685]: I1204 06:26:01.454098 4685 generic.go:334] "Generic (PLEG): container finished" podID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerID="23290284a93ea74caa6b6721ba921f3609874cca77717c1991824e02ce44378a" exitCode=0 Dec 04 06:26:01 crc kubenswrapper[4685]: I1204 06:26:01.454179 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f58c7" event={"ID":"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f","Type":"ContainerDied","Data":"23290284a93ea74caa6b6721ba921f3609874cca77717c1991824e02ce44378a"} Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.530762 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.636999 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l78df\" (UniqueName: \"kubernetes.io/projected/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-kube-api-access-l78df\") pod \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.637143 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-catalog-content\") pod \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.637216 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-utilities\") pod \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\" (UID: \"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f\") " Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.639343 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-utilities" (OuterVolumeSpecName: "utilities") pod "c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" (UID: "c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.665367 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-kube-api-access-l78df" (OuterVolumeSpecName: "kube-api-access-l78df") pod "c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" (UID: "c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f"). InnerVolumeSpecName "kube-api-access-l78df". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.687868 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" (UID: "c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.739617 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.740065 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:26:02 crc kubenswrapper[4685]: I1204 06:26:02.740085 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l78df\" (UniqueName: \"kubernetes.io/projected/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f-kube-api-access-l78df\") on node \"crc\" DevicePath \"\"" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.472733 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" event={"ID":"90f5cdd3-8605-4cb5-9aea-83d57954bc17","Type":"ContainerStarted","Data":"e9b59728d02ea39d5b5316f196a1d93ebec6f6d61fab34e29315c8ee3aa8e778"} Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.472882 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.476255 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f58c7" event={"ID":"c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f","Type":"ContainerDied","Data":"6fbd2b9b8b93b39c238e553506726522519fa1498016aa9e36733352d4619d7a"} Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.476328 4685 scope.go:117] "RemoveContainer" containerID="23290284a93ea74caa6b6721ba921f3609874cca77717c1991824e02ce44378a" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.476517 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f58c7" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.501067 4685 scope.go:117] "RemoveContainer" containerID="8a9ac670a1242d05faa3655161c1fdc538ecbf7b68df30cf416994dfeaa07bee" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.527677 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" podStartSLOduration=1.910983425 podStartE2EDuration="7.527108941s" podCreationTimestamp="2025-12-04 06:25:56 +0000 UTC" firstStartedPulling="2025-12-04 06:25:56.990045297 +0000 UTC m=+894.178276072" lastFinishedPulling="2025-12-04 06:26:02.606170803 +0000 UTC m=+899.794401588" observedRunningTime="2025-12-04 06:26:03.519671513 +0000 UTC m=+900.707902308" watchObservedRunningTime="2025-12-04 06:26:03.527108941 +0000 UTC m=+900.715339716" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.543272 4685 scope.go:117] "RemoveContainer" containerID="c39cc1f94aca01975d2321d68386733057d8442a64e9255c858794ab746ed5d6" Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.545974 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f58c7"] Dec 04 06:26:03 crc kubenswrapper[4685]: I1204 06:26:03.554062 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f58c7"] Dec 04 06:26:05 crc kubenswrapper[4685]: I1204 06:26:05.134565 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" path="/var/lib/kubelet/pods/c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f/volumes" Dec 04 06:26:06 crc kubenswrapper[4685]: I1204 06:26:06.121530 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:26:06 crc kubenswrapper[4685]: I1204 06:26:06.121995 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:26:16 crc kubenswrapper[4685]: I1204 06:26:16.705758 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-778b7fc6d-66vch" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.023590 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl"] Dec 04 06:26:36 crc kubenswrapper[4685]: E1204 06:26:36.024935 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="registry-server" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.024959 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="registry-server" Dec 04 06:26:36 crc kubenswrapper[4685]: E1204 06:26:36.024985 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="extract-utilities" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.024995 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="extract-utilities" Dec 04 06:26:36 crc kubenswrapper[4685]: E1204 06:26:36.025017 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="extract-content" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.025027 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="extract-content" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.025182 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1ffb52e-d3a8-4a40-b9d9-dc5414534a1f" containerName="registry-server" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.026156 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.028964 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-pddsd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.035559 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.052695 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.055704 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.058614 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4wl7\" (UniqueName: \"kubernetes.io/projected/c9f1a1e0-3fd0-4c69-8475-9ed056330d57-kube-api-access-m4wl7\") pod \"barbican-operator-controller-manager-7d9dfd778-h8fbl\" (UID: \"c9f1a1e0-3fd0-4c69-8475-9ed056330d57\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.063755 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-749gg" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.068659 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.092269 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.093950 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.097479 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.110054 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-vnlsx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.119207 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.120317 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.122742 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.122819 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7hxvb" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.122835 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.130701 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.142636 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.144054 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.148694 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-hhzvg" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.164029 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4mgh\" (UniqueName: \"kubernetes.io/projected/7aa55d54-0ab9-4b25-8226-095d3703ca03-kube-api-access-h4mgh\") pod \"designate-operator-controller-manager-78b4bc895b-6pgsq\" (UID: \"7aa55d54-0ab9-4b25-8226-095d3703ca03\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.164264 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4wl7\" (UniqueName: \"kubernetes.io/projected/c9f1a1e0-3fd0-4c69-8475-9ed056330d57-kube-api-access-m4wl7\") pod \"barbican-operator-controller-manager-7d9dfd778-h8fbl\" (UID: \"c9f1a1e0-3fd0-4c69-8475-9ed056330d57\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.164329 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p5wb\" (UniqueName: \"kubernetes.io/projected/87b87b6f-538b-4ba3-ba4c-994924ffe76e-kube-api-access-2p5wb\") pod \"heat-operator-controller-manager-5f64f6f8bb-rt2mx\" (UID: \"87b87b6f-538b-4ba3-ba4c-994924ffe76e\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.164570 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cdws\" (UniqueName: \"kubernetes.io/projected/4c3ba289-3038-4c41-881a-d53e4f54bc47-kube-api-access-7cdws\") pod \"cinder-operator-controller-manager-859b6ccc6-s8zzd\" (UID: \"4c3ba289-3038-4c41-881a-d53e4f54bc47\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.164648 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzql4\" (UniqueName: \"kubernetes.io/projected/231797d5-080c-4e7f-ad27-186a01bf4094-kube-api-access-xzql4\") pod \"glance-operator-controller-manager-77987cd8cd-czngd\" (UID: \"231797d5-080c-4e7f-ad27-186a01bf4094\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.180837 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.195022 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.196125 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.201774 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-fkhqx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.241081 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.248850 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4wl7\" (UniqueName: \"kubernetes.io/projected/c9f1a1e0-3fd0-4c69-8475-9ed056330d57-kube-api-access-m4wl7\") pod \"barbican-operator-controller-manager-7d9dfd778-h8fbl\" (UID: \"c9f1a1e0-3fd0-4c69-8475-9ed056330d57\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.255823 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.257021 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.258944 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9djw5" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.263081 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.268089 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p5wb\" (UniqueName: \"kubernetes.io/projected/87b87b6f-538b-4ba3-ba4c-994924ffe76e-kube-api-access-2p5wb\") pod \"heat-operator-controller-manager-5f64f6f8bb-rt2mx\" (UID: \"87b87b6f-538b-4ba3-ba4c-994924ffe76e\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.268190 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cdws\" (UniqueName: \"kubernetes.io/projected/4c3ba289-3038-4c41-881a-d53e4f54bc47-kube-api-access-7cdws\") pod \"cinder-operator-controller-manager-859b6ccc6-s8zzd\" (UID: \"4c3ba289-3038-4c41-881a-d53e4f54bc47\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.268221 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzql4\" (UniqueName: \"kubernetes.io/projected/231797d5-080c-4e7f-ad27-186a01bf4094-kube-api-access-xzql4\") pod \"glance-operator-controller-manager-77987cd8cd-czngd\" (UID: \"231797d5-080c-4e7f-ad27-186a01bf4094\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.268259 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4mgh\" (UniqueName: \"kubernetes.io/projected/7aa55d54-0ab9-4b25-8226-095d3703ca03-kube-api-access-h4mgh\") pod \"designate-operator-controller-manager-78b4bc895b-6pgsq\" (UID: \"7aa55d54-0ab9-4b25-8226-095d3703ca03\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.268296 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2hcj\" (UniqueName: \"kubernetes.io/projected/211501da-f7e5-405e-b1cb-2673ca2b82c1-kube-api-access-t2hcj\") pod \"horizon-operator-controller-manager-68c6d99b8f-wd5jl\" (UID: \"211501da-f7e5-405e-b1cb-2673ca2b82c1\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.278127 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.279092 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.281358 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-9p29l" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.291445 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.292496 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.299144 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tgbmm" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.300188 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p5wb\" (UniqueName: \"kubernetes.io/projected/87b87b6f-538b-4ba3-ba4c-994924ffe76e-kube-api-access-2p5wb\") pod \"heat-operator-controller-manager-5f64f6f8bb-rt2mx\" (UID: \"87b87b6f-538b-4ba3-ba4c-994924ffe76e\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.311783 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.324199 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzql4\" (UniqueName: \"kubernetes.io/projected/231797d5-080c-4e7f-ad27-186a01bf4094-kube-api-access-xzql4\") pod \"glance-operator-controller-manager-77987cd8cd-czngd\" (UID: \"231797d5-080c-4e7f-ad27-186a01bf4094\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.327578 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cdws\" (UniqueName: \"kubernetes.io/projected/4c3ba289-3038-4c41-881a-d53e4f54bc47-kube-api-access-7cdws\") pod \"cinder-operator-controller-manager-859b6ccc6-s8zzd\" (UID: \"4c3ba289-3038-4c41-881a-d53e4f54bc47\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.352032 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4mgh\" (UniqueName: \"kubernetes.io/projected/7aa55d54-0ab9-4b25-8226-095d3703ca03-kube-api-access-h4mgh\") pod \"designate-operator-controller-manager-78b4bc895b-6pgsq\" (UID: \"7aa55d54-0ab9-4b25-8226-095d3703ca03\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.365500 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.369804 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvb64\" (UniqueName: \"kubernetes.io/projected/cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1-kube-api-access-qvb64\") pod \"keystone-operator-controller-manager-7765d96ddf-xql6q\" (UID: \"cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.369859 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.369882 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fkk7\" (UniqueName: \"kubernetes.io/projected/d55943b8-039d-4747-b341-d0fe2e0f6df7-kube-api-access-9fkk7\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.369900 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmp9l\" (UniqueName: \"kubernetes.io/projected/ee4e7fd4-46c8-4aac-be76-5097d9f76ac2-kube-api-access-tmp9l\") pod \"ironic-operator-controller-manager-6c548fd776-qqd9p\" (UID: \"ee4e7fd4-46c8-4aac-be76-5097d9f76ac2\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.369939 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2hcj\" (UniqueName: \"kubernetes.io/projected/211501da-f7e5-405e-b1cb-2673ca2b82c1-kube-api-access-t2hcj\") pod \"horizon-operator-controller-manager-68c6d99b8f-wd5jl\" (UID: \"211501da-f7e5-405e-b1cb-2673ca2b82c1\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.397773 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.410640 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.416134 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.421848 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.425951 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.426171 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2hcj\" (UniqueName: \"kubernetes.io/projected/211501da-f7e5-405e-b1cb-2673ca2b82c1-kube-api-access-t2hcj\") pod \"horizon-operator-controller-manager-68c6d99b8f-wd5jl\" (UID: \"211501da-f7e5-405e-b1cb-2673ca2b82c1\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.426944 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.429684 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.438709 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.438788 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.439337 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2gt2w" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.441623 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-95wcn" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.445020 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.446562 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.448568 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.455424 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.456922 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.457354 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2hbf9" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.466784 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.466998 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.468157 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.470017 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-6bxjj" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.470748 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.471925 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.471960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fkk7\" (UniqueName: \"kubernetes.io/projected/d55943b8-039d-4747-b341-d0fe2e0f6df7-kube-api-access-9fkk7\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.471983 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmp9l\" (UniqueName: \"kubernetes.io/projected/ee4e7fd4-46c8-4aac-be76-5097d9f76ac2-kube-api-access-tmp9l\") pod \"ironic-operator-controller-manager-6c548fd776-qqd9p\" (UID: \"ee4e7fd4-46c8-4aac-be76-5097d9f76ac2\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.472060 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvb64\" (UniqueName: \"kubernetes.io/projected/cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1-kube-api-access-qvb64\") pod \"keystone-operator-controller-manager-7765d96ddf-xql6q\" (UID: \"cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:26:36 crc kubenswrapper[4685]: E1204 06:26:36.472349 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:36 crc kubenswrapper[4685]: E1204 06:26:36.472388 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert podName:d55943b8-039d-4747-b341-d0fe2e0f6df7 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:36.972374431 +0000 UTC m=+934.160605206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert") pod "infra-operator-controller-manager-57548d458d-8fjlw" (UID: "d55943b8-039d-4747-b341-d0fe2e0f6df7") : secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.483847 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-mt9wv" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.490006 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.492696 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.505359 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvb64\" (UniqueName: \"kubernetes.io/projected/cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1-kube-api-access-qvb64\") pod \"keystone-operator-controller-manager-7765d96ddf-xql6q\" (UID: \"cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.505456 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fkk7\" (UniqueName: \"kubernetes.io/projected/d55943b8-039d-4747-b341-d0fe2e0f6df7-kube-api-access-9fkk7\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.505804 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmp9l\" (UniqueName: \"kubernetes.io/projected/ee4e7fd4-46c8-4aac-be76-5097d9f76ac2-kube-api-access-tmp9l\") pod \"ironic-operator-controller-manager-6c548fd776-qqd9p\" (UID: \"ee4e7fd4-46c8-4aac-be76-5097d9f76ac2\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.546277 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.553220 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.580483 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97zmw\" (UniqueName: \"kubernetes.io/projected/89fc2488-fa6d-4515-923e-115aa45e23d4-kube-api-access-97zmw\") pod \"mariadb-operator-controller-manager-56bbcc9d85-dnkjj\" (UID: \"89fc2488-fa6d-4515-923e-115aa45e23d4\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.580587 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zghz\" (UniqueName: \"kubernetes.io/projected/4af99755-81d3-43ec-a924-a3dcc4347315-kube-api-access-2zghz\") pod \"manila-operator-controller-manager-7c79b5df47-rrk46\" (UID: \"4af99755-81d3-43ec-a924-a3dcc4347315\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.580624 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlmp2\" (UniqueName: \"kubernetes.io/projected/e9b56e59-1bc8-4608-bdfe-2901e30c7009-kube-api-access-qlmp2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-7t9xh\" (UID: \"e9b56e59-1bc8-4608-bdfe-2901e30c7009\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.580670 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66krq\" (UniqueName: \"kubernetes.io/projected/c7545a63-4651-4d38-a251-9de38ce5a226-kube-api-access-66krq\") pod \"nova-operator-controller-manager-697bc559fc-jlm87\" (UID: \"c7545a63-4651-4d38-a251-9de38ce5a226\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.675193 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.676371 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.680161 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-gw8l9" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.681550 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97zmw\" (UniqueName: \"kubernetes.io/projected/89fc2488-fa6d-4515-923e-115aa45e23d4-kube-api-access-97zmw\") pod \"mariadb-operator-controller-manager-56bbcc9d85-dnkjj\" (UID: \"89fc2488-fa6d-4515-923e-115aa45e23d4\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.681609 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zghz\" (UniqueName: \"kubernetes.io/projected/4af99755-81d3-43ec-a924-a3dcc4347315-kube-api-access-2zghz\") pod \"manila-operator-controller-manager-7c79b5df47-rrk46\" (UID: \"4af99755-81d3-43ec-a924-a3dcc4347315\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.681630 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlmp2\" (UniqueName: \"kubernetes.io/projected/e9b56e59-1bc8-4608-bdfe-2901e30c7009-kube-api-access-qlmp2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-7t9xh\" (UID: \"e9b56e59-1bc8-4608-bdfe-2901e30c7009\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.681658 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66krq\" (UniqueName: \"kubernetes.io/projected/c7545a63-4651-4d38-a251-9de38ce5a226-kube-api-access-66krq\") pod \"nova-operator-controller-manager-697bc559fc-jlm87\" (UID: \"c7545a63-4651-4d38-a251-9de38ce5a226\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.681687 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-948tx\" (UniqueName: \"kubernetes.io/projected/09248202-ba20-4b70-95ab-1a7e67c04ef7-kube-api-access-948tx\") pod \"octavia-operator-controller-manager-998648c74-v2lxd\" (UID: \"09248202-ba20-4b70-95ab-1a7e67c04ef7\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.694066 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nhb96"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.695201 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.703571 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7npkx" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.730776 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97zmw\" (UniqueName: \"kubernetes.io/projected/89fc2488-fa6d-4515-923e-115aa45e23d4-kube-api-access-97zmw\") pod \"mariadb-operator-controller-manager-56bbcc9d85-dnkjj\" (UID: \"89fc2488-fa6d-4515-923e-115aa45e23d4\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.731071 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.742254 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zghz\" (UniqueName: \"kubernetes.io/projected/4af99755-81d3-43ec-a924-a3dcc4347315-kube-api-access-2zghz\") pod \"manila-operator-controller-manager-7c79b5df47-rrk46\" (UID: \"4af99755-81d3-43ec-a924-a3dcc4347315\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.753505 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.755062 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.760079 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fxcb7" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.765242 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.765770 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.774116 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlmp2\" (UniqueName: \"kubernetes.io/projected/e9b56e59-1bc8-4608-bdfe-2901e30c7009-kube-api-access-qlmp2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-7t9xh\" (UID: \"e9b56e59-1bc8-4608-bdfe-2901e30c7009\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.775035 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66krq\" (UniqueName: \"kubernetes.io/projected/c7545a63-4651-4d38-a251-9de38ce5a226-kube-api-access-66krq\") pod \"nova-operator-controller-manager-697bc559fc-jlm87\" (UID: \"c7545a63-4651-4d38-a251-9de38ce5a226\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.779949 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.785863 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-948tx\" (UniqueName: \"kubernetes.io/projected/09248202-ba20-4b70-95ab-1a7e67c04ef7-kube-api-access-948tx\") pod \"octavia-operator-controller-manager-998648c74-v2lxd\" (UID: \"09248202-ba20-4b70-95ab-1a7e67c04ef7\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.785944 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnnv8\" (UniqueName: \"kubernetes.io/projected/efab633f-9698-443f-8175-279a82810d4d-kube-api-access-qnnv8\") pod \"ovn-operator-controller-manager-b6456fdb6-cdbkv\" (UID: \"efab633f-9698-443f-8175-279a82810d4d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.807745 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.811435 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-948tx\" (UniqueName: \"kubernetes.io/projected/09248202-ba20-4b70-95ab-1a7e67c04ef7-kube-api-access-948tx\") pod \"octavia-operator-controller-manager-998648c74-v2lxd\" (UID: \"09248202-ba20-4b70-95ab-1a7e67c04ef7\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.842467 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.870967 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.879967 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.881598 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.890133 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.890190 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnnv8\" (UniqueName: \"kubernetes.io/projected/efab633f-9698-443f-8175-279a82810d4d-kube-api-access-qnnv8\") pod \"ovn-operator-controller-manager-b6456fdb6-cdbkv\" (UID: \"efab633f-9698-443f-8175-279a82810d4d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.890215 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cbkq\" (UniqueName: \"kubernetes.io/projected/18233d1f-7861-4bd5-8643-f464fce185ca-kube-api-access-6cbkq\") pod \"swift-operator-controller-manager-5f8c65bbfc-xchfb\" (UID: \"18233d1f-7861-4bd5-8643-f464fce185ca\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.890257 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t76j8\" (UniqueName: \"kubernetes.io/projected/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-kube-api-access-t76j8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.890304 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwtz7\" (UniqueName: \"kubernetes.io/projected/e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13-kube-api-access-hwtz7\") pod \"placement-operator-controller-manager-78f8948974-nhb96\" (UID: \"e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.907735 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-7hw28" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.908622 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.909147 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.910608 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.928545 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-v4m96" Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.929190 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nhb96"] Dec 04 06:26:36 crc kubenswrapper[4685]: I1204 06:26:36.951073 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnnv8\" (UniqueName: \"kubernetes.io/projected/efab633f-9698-443f-8175-279a82810d4d-kube-api-access-qnnv8\") pod \"ovn-operator-controller-manager-b6456fdb6-cdbkv\" (UID: \"efab633f-9698-443f-8175-279a82810d4d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.014257 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.027885 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.060307 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwtz7\" (UniqueName: \"kubernetes.io/projected/e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13-kube-api-access-hwtz7\") pod \"placement-operator-controller-manager-78f8948974-nhb96\" (UID: \"e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.060386 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.060435 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2mbh\" (UniqueName: \"kubernetes.io/projected/9cee5367-7803-44a8-b047-66e4311f9e0a-kube-api-access-v2mbh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-c64mf\" (UID: \"9cee5367-7803-44a8-b047-66e4311f9e0a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.060485 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cbkq\" (UniqueName: \"kubernetes.io/projected/18233d1f-7861-4bd5-8643-f464fce185ca-kube-api-access-6cbkq\") pod \"swift-operator-controller-manager-5f8c65bbfc-xchfb\" (UID: \"18233d1f-7861-4bd5-8643-f464fce185ca\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.060534 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.060560 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t76j8\" (UniqueName: \"kubernetes.io/projected/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-kube-api-access-t76j8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.076056 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.076131 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert podName:3ed35c5e-42ad-433c-a044-0fdac4c8ad06 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:37.576109303 +0000 UTC m=+934.764340078 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" (UID: "3ed35c5e-42ad-433c-a044-0fdac4c8ad06") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.076651 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.076697 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert podName:d55943b8-039d-4747-b341-d0fe2e0f6df7 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:38.076682462 +0000 UTC m=+935.264913237 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert") pod "infra-operator-controller-manager-57548d458d-8fjlw" (UID: "d55943b8-039d-4747-b341-d0fe2e0f6df7") : secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.083693 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.165556 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2mbh\" (UniqueName: \"kubernetes.io/projected/9cee5367-7803-44a8-b047-66e4311f9e0a-kube-api-access-v2mbh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-c64mf\" (UID: \"9cee5367-7803-44a8-b047-66e4311f9e0a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.171741 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.176531 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t76j8\" (UniqueName: \"kubernetes.io/projected/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-kube-api-access-t76j8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.177320 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cbkq\" (UniqueName: \"kubernetes.io/projected/18233d1f-7861-4bd5-8643-f464fce185ca-kube-api-access-6cbkq\") pod \"swift-operator-controller-manager-5f8c65bbfc-xchfb\" (UID: \"18233d1f-7861-4bd5-8643-f464fce185ca\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.189435 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.200557 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2mbh\" (UniqueName: \"kubernetes.io/projected/9cee5367-7803-44a8-b047-66e4311f9e0a-kube-api-access-v2mbh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-c64mf\" (UID: \"9cee5367-7803-44a8-b047-66e4311f9e0a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.250668 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwtz7\" (UniqueName: \"kubernetes.io/projected/e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13-kube-api-access-hwtz7\") pod \"placement-operator-controller-manager-78f8948974-nhb96\" (UID: \"e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.250817 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.256099 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.262080 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-9cjnh" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.262801 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.276956 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx59q\" (UniqueName: \"kubernetes.io/projected/e3ac8015-c6a6-4837-9118-0de5f1b351df-kube-api-access-dx59q\") pod \"test-operator-controller-manager-5854674fcc-x7wr8\" (UID: \"e3ac8015-c6a6-4837-9118-0de5f1b351df\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.335239 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.369128 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.385530 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx59q\" (UniqueName: \"kubernetes.io/projected/e3ac8015-c6a6-4837-9118-0de5f1b351df-kube-api-access-dx59q\") pod \"test-operator-controller-manager-5854674fcc-x7wr8\" (UID: \"e3ac8015-c6a6-4837-9118-0de5f1b351df\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.394096 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.405368 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.416519 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lxq7n" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.422676 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx59q\" (UniqueName: \"kubernetes.io/projected/e3ac8015-c6a6-4837-9118-0de5f1b351df-kube-api-access-dx59q\") pod \"test-operator-controller-manager-5854674fcc-x7wr8\" (UID: \"e3ac8015-c6a6-4837-9118-0de5f1b351df\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.424523 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.445301 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.492568 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.546653 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.548032 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.552641 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.552821 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.554883 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-f7csq" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.568044 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.578661 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.581402 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.584529 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-h57ch" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.592554 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.592706 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vhpl\" (UniqueName: \"kubernetes.io/projected/639879a9-dc01-4145-916b-8d568130af06-kube-api-access-6vhpl\") pod \"watcher-operator-controller-manager-769dc69bc-zzn46\" (UID: \"639879a9-dc01-4145-916b-8d568130af06\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.592968 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.593033 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert podName:3ed35c5e-42ad-433c-a044-0fdac4c8ad06 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:38.593011212 +0000 UTC m=+935.781241987 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" (UID: "3ed35c5e-42ad-433c-a044-0fdac4c8ad06") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.593916 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.593978 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.603692 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.620287 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.696094 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.696228 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vhpl\" (UniqueName: \"kubernetes.io/projected/639879a9-dc01-4145-916b-8d568130af06-kube-api-access-6vhpl\") pod \"watcher-operator-controller-manager-769dc69bc-zzn46\" (UID: \"639879a9-dc01-4145-916b-8d568130af06\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.696264 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjjd7\" (UniqueName: \"kubernetes.io/projected/541675b0-83d5-49cf-8209-2de2d8385f3b-kube-api-access-xjjd7\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.696301 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.696356 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdrvm\" (UniqueName: \"kubernetes.io/projected/ecb2e435-29cf-4b37-a0bb-4f2441cab009-kube-api-access-bdrvm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tmmg4\" (UID: \"ecb2e435-29cf-4b37-a0bb-4f2441cab009\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.727287 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vhpl\" (UniqueName: \"kubernetes.io/projected/639879a9-dc01-4145-916b-8d568130af06-kube-api-access-6vhpl\") pod \"watcher-operator-controller-manager-769dc69bc-zzn46\" (UID: \"639879a9-dc01-4145-916b-8d568130af06\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.747185 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.749895 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.790857 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" event={"ID":"231797d5-080c-4e7f-ad27-186a01bf4094","Type":"ContainerStarted","Data":"6248eae048284fcab09e009d6f45cdf51905dde1f6637ee897f4010ea4243622"} Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.792114 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" event={"ID":"c9f1a1e0-3fd0-4c69-8475-9ed056330d57","Type":"ContainerStarted","Data":"07bd17f371313c8a2a826b3b8f1e9221581b8eb5ef3bbb4324912e6d028a2e27"} Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.797303 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdrvm\" (UniqueName: \"kubernetes.io/projected/ecb2e435-29cf-4b37-a0bb-4f2441cab009-kube-api-access-bdrvm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tmmg4\" (UID: \"ecb2e435-29cf-4b37-a0bb-4f2441cab009\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.797392 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.797477 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjjd7\" (UniqueName: \"kubernetes.io/projected/541675b0-83d5-49cf-8209-2de2d8385f3b-kube-api-access-xjjd7\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.797506 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.797643 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.797658 4685 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.797768 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:38.297731717 +0000 UTC m=+935.485962692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "webhook-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: E1204 06:26:37.797826 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:38.297798669 +0000 UTC m=+935.486029434 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "metrics-server-cert" not found Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.799551 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" event={"ID":"4c3ba289-3038-4c41-881a-d53e4f54bc47","Type":"ContainerStarted","Data":"dc2336529b9d7bd81c588125140784c7e8c04fe209289ed595d5bdbcbc132291"} Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.838086 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl"] Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.846894 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdrvm\" (UniqueName: \"kubernetes.io/projected/ecb2e435-29cf-4b37-a0bb-4f2441cab009-kube-api-access-bdrvm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tmmg4\" (UID: \"ecb2e435-29cf-4b37-a0bb-4f2441cab009\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.850931 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjjd7\" (UniqueName: \"kubernetes.io/projected/541675b0-83d5-49cf-8209-2de2d8385f3b-kube-api-access-xjjd7\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:37 crc kubenswrapper[4685]: I1204 06:26:37.977431 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.102317 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.102576 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.102623 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert podName:d55943b8-039d-4747-b341-d0fe2e0f6df7 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:40.102610078 +0000 UTC m=+937.290840853 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert") pod "infra-operator-controller-manager-57548d458d-8fjlw" (UID: "d55943b8-039d-4747-b341-d0fe2e0f6df7") : secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.300526 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.306602 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.306769 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.306813 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.306889 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:39.306870518 +0000 UTC m=+936.495101293 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "webhook-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.310147 4685 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.310373 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:39.310340389 +0000 UTC m=+936.498571344 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "metrics-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.315969 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.322004 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p"] Dec 04 06:26:38 crc kubenswrapper[4685]: W1204 06:26:38.322711 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9b56e59_1bc8_4608_bdfe_2901e30c7009.slice/crio-29f92c2adb3df29d53b1998035d36e860840816b8195da7c055a7b95cd93c7ac WatchSource:0}: Error finding container 29f92c2adb3df29d53b1998035d36e860840816b8195da7c055a7b95cd93c7ac: Status 404 returned error can't find the container with id 29f92c2adb3df29d53b1998035d36e860840816b8195da7c055a7b95cd93c7ac Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.335768 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.340617 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.614495 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.614684 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.614732 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert podName:3ed35c5e-42ad-433c-a044-0fdac4c8ad06 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:40.614717245 +0000 UTC m=+937.802948020 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" (UID: "3ed35c5e-42ad-433c-a044-0fdac4c8ad06") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.684542 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.730177 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.746478 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.751519 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.758279 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.763270 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8"] Dec 04 06:26:38 crc kubenswrapper[4685]: W1204 06:26:38.763744 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7794d4a_b9b0_4d3c_995f_98f8bd2dbd13.slice/crio-1ca1c97c63c4f287465d9f16866fc23e8d70a94735d91bb03f777a8d74fe305e WatchSource:0}: Error finding container 1ca1c97c63c4f287465d9f16866fc23e8d70a94735d91bb03f777a8d74fe305e: Status 404 returned error can't find the container with id 1ca1c97c63c4f287465d9f16866fc23e8d70a94735d91bb03f777a8d74fe305e Dec 04 06:26:38 crc kubenswrapper[4685]: W1204 06:26:38.771339 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3ac8015_c6a6_4837_9118_0de5f1b351df.slice/crio-ef83f7df1cc8fa490d7ed7cbbb6f4ed5a796da5134eeb6f89600d95b772c34db WatchSource:0}: Error finding container ef83f7df1cc8fa490d7ed7cbbb6f4ed5a796da5134eeb6f89600d95b772c34db: Status 404 returned error can't find the container with id ef83f7df1cc8fa490d7ed7cbbb6f4ed5a796da5134eeb6f89600d95b772c34db Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.774926 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb"] Dec 04 06:26:38 crc kubenswrapper[4685]: W1204 06:26:38.778472 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecb2e435_29cf_4b37_a0bb_4f2441cab009.slice/crio-53f73c8be5e52ff78fa34b849a66cf7d29369fbbb571f308d47be21e40ff50a6 WatchSource:0}: Error finding container 53f73c8be5e52ff78fa34b849a66cf7d29369fbbb571f308d47be21e40ff50a6: Status 404 returned error can't find the container with id 53f73c8be5e52ff78fa34b849a66cf7d29369fbbb571f308d47be21e40ff50a6 Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.785702 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nhb96"] Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.791024 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bdrvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-tmmg4_openstack-operators(ecb2e435-29cf-4b37-a0bb-4f2441cab009): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.791020 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dx59q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-x7wr8_openstack-operators(e3ac8015-c6a6-4837-9118-0de5f1b351df): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.792226 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" podUID="ecb2e435-29cf-4b37-a0bb-4f2441cab009" Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.795961 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf"] Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.796443 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dx59q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-x7wr8_openstack-operators(e3ac8015-c6a6-4837-9118-0de5f1b351df): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.796459 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hwtz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-nhb96_openstack-operators(e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.797725 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" podUID="e3ac8015-c6a6-4837-9118-0de5f1b351df" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.799212 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hwtz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-nhb96_openstack-operators(e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.800429 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" podUID="e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13" Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.801597 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4"] Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.816094 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" event={"ID":"cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1","Type":"ContainerStarted","Data":"6f2f335a7e1d5507ca4de574ee9ab9a4f592e77664c865b398b2680c2b839b6f"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.819924 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" event={"ID":"09248202-ba20-4b70-95ab-1a7e67c04ef7","Type":"ContainerStarted","Data":"07485c42d24fbd79e2d23409e4f4deafb6a1f7dfadf20f6c2b8924923a8e2831"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.822859 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" event={"ID":"4af99755-81d3-43ec-a924-a3dcc4347315","Type":"ContainerStarted","Data":"17a336396912df0b47e37dbf2f009f18536f085fc3c9184a0877d9016cadb05e"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.827221 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" event={"ID":"89fc2488-fa6d-4515-923e-115aa45e23d4","Type":"ContainerStarted","Data":"8ec2bf54b82f14a0e79a513461adc79e3b2dffcc60100ec31b87117bae228323"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.830276 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" event={"ID":"87b87b6f-538b-4ba3-ba4c-994924ffe76e","Type":"ContainerStarted","Data":"05f9a55a114335de01e054762a88e7ccbd7c1bc7250b83e242c9ddc91ebbfc97"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.832200 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" event={"ID":"9cee5367-7803-44a8-b047-66e4311f9e0a","Type":"ContainerStarted","Data":"782e59a14549f78a11b4145158bc71b740da1e7d0292a8d56e8dc9ac61690e33"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.833550 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" event={"ID":"211501da-f7e5-405e-b1cb-2673ca2b82c1","Type":"ContainerStarted","Data":"93943aea694c2aa288204b411e297c3192cb916e88073fd35804b9c1b122ad8d"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.836046 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" event={"ID":"c7545a63-4651-4d38-a251-9de38ce5a226","Type":"ContainerStarted","Data":"d98a4357e6137c995feedef65b3fbf84fa68a129af16b56549500cba0fce4564"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.845166 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" event={"ID":"e9b56e59-1bc8-4608-bdfe-2901e30c7009","Type":"ContainerStarted","Data":"29f92c2adb3df29d53b1998035d36e860840816b8195da7c055a7b95cd93c7ac"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.850574 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" event={"ID":"e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13","Type":"ContainerStarted","Data":"1ca1c97c63c4f287465d9f16866fc23e8d70a94735d91bb03f777a8d74fe305e"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.855602 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" event={"ID":"e3ac8015-c6a6-4837-9118-0de5f1b351df","Type":"ContainerStarted","Data":"ef83f7df1cc8fa490d7ed7cbbb6f4ed5a796da5134eeb6f89600d95b772c34db"} Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.857508 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" podUID="e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13" Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.857661 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" podUID="e3ac8015-c6a6-4837-9118-0de5f1b351df" Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.858565 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" event={"ID":"ee4e7fd4-46c8-4aac-be76-5097d9f76ac2","Type":"ContainerStarted","Data":"ccdd51785a9e5248ff13ef2c9b9db29ceb30b3d58f233733c3cbbbca7ed85886"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.860280 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" event={"ID":"639879a9-dc01-4145-916b-8d568130af06","Type":"ContainerStarted","Data":"cb5253f9625c5fb78c39752923c476f1394e1415b5ab833ef0ecf53cbb07a093"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.861596 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" event={"ID":"efab633f-9698-443f-8175-279a82810d4d","Type":"ContainerStarted","Data":"73abec75db1f0ed43a9825e36dac101f5603a2dd26b9c1e8faa1a64d61bf42c4"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.863357 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" event={"ID":"18233d1f-7861-4bd5-8643-f464fce185ca","Type":"ContainerStarted","Data":"ab8385de14427342cd34f87012cae944162eefe9d2c079ce4dff807bc2a0567c"} Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.864514 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" event={"ID":"ecb2e435-29cf-4b37-a0bb-4f2441cab009","Type":"ContainerStarted","Data":"53f73c8be5e52ff78fa34b849a66cf7d29369fbbb571f308d47be21e40ff50a6"} Dec 04 06:26:38 crc kubenswrapper[4685]: E1204 06:26:38.866484 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" podUID="ecb2e435-29cf-4b37-a0bb-4f2441cab009" Dec 04 06:26:38 crc kubenswrapper[4685]: I1204 06:26:38.866811 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" event={"ID":"7aa55d54-0ab9-4b25-8226-095d3703ca03","Type":"ContainerStarted","Data":"fa53149efd45a522ed2b4c2a57ae9bda7b9fb1f429dc650c133b7adf68225ae0"} Dec 04 06:26:39 crc kubenswrapper[4685]: I1204 06:26:39.335179 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.335593 4685 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 06:26:39 crc kubenswrapper[4685]: I1204 06:26:39.335615 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.335675 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:41.335653537 +0000 UTC m=+938.523884372 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "metrics-server-cert" not found Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.335773 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.335824 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:41.335806822 +0000 UTC m=+938.524037597 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "webhook-server-cert" not found Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.881184 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" podUID="ecb2e435-29cf-4b37-a0bb-4f2441cab009" Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.882864 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" podUID="e3ac8015-c6a6-4837-9118-0de5f1b351df" Dec 04 06:26:39 crc kubenswrapper[4685]: E1204 06:26:39.883385 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" podUID="e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13" Dec 04 06:26:40 crc kubenswrapper[4685]: I1204 06:26:40.148883 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:40 crc kubenswrapper[4685]: E1204 06:26:40.149085 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:40 crc kubenswrapper[4685]: E1204 06:26:40.149162 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert podName:d55943b8-039d-4747-b341-d0fe2e0f6df7 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:44.149138203 +0000 UTC m=+941.337368978 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert") pod "infra-operator-controller-manager-57548d458d-8fjlw" (UID: "d55943b8-039d-4747-b341-d0fe2e0f6df7") : secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:40 crc kubenswrapper[4685]: I1204 06:26:40.656961 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:40 crc kubenswrapper[4685]: E1204 06:26:40.657214 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:40 crc kubenswrapper[4685]: E1204 06:26:40.657280 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert podName:3ed35c5e-42ad-433c-a044-0fdac4c8ad06 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:44.657258013 +0000 UTC m=+941.845488788 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" (UID: "3ed35c5e-42ad-433c-a044-0fdac4c8ad06") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:41 crc kubenswrapper[4685]: I1204 06:26:41.372878 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:41 crc kubenswrapper[4685]: I1204 06:26:41.372973 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:41 crc kubenswrapper[4685]: E1204 06:26:41.373132 4685 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 06:26:41 crc kubenswrapper[4685]: E1204 06:26:41.373184 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:45.373169095 +0000 UTC m=+942.561399870 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "metrics-server-cert" not found Dec 04 06:26:41 crc kubenswrapper[4685]: E1204 06:26:41.373698 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 06:26:41 crc kubenswrapper[4685]: E1204 06:26:41.374445 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:45.374390444 +0000 UTC m=+942.562621219 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "webhook-server-cert" not found Dec 04 06:26:44 crc kubenswrapper[4685]: I1204 06:26:44.152775 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:44 crc kubenswrapper[4685]: E1204 06:26:44.153073 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:44 crc kubenswrapper[4685]: E1204 06:26:44.153390 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert podName:d55943b8-039d-4747-b341-d0fe2e0f6df7 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:52.153364017 +0000 UTC m=+949.341594792 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert") pod "infra-operator-controller-manager-57548d458d-8fjlw" (UID: "d55943b8-039d-4747-b341-d0fe2e0f6df7") : secret "infra-operator-webhook-server-cert" not found Dec 04 06:26:44 crc kubenswrapper[4685]: I1204 06:26:44.662568 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:44 crc kubenswrapper[4685]: E1204 06:26:44.662864 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:44 crc kubenswrapper[4685]: E1204 06:26:44.662940 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert podName:3ed35c5e-42ad-433c-a044-0fdac4c8ad06 nodeName:}" failed. No retries permitted until 2025-12-04 06:26:52.662919212 +0000 UTC m=+949.851149987 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" (UID: "3ed35c5e-42ad-433c-a044-0fdac4c8ad06") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 06:26:45 crc kubenswrapper[4685]: I1204 06:26:45.376358 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:45 crc kubenswrapper[4685]: E1204 06:26:45.376646 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 06:26:45 crc kubenswrapper[4685]: I1204 06:26:45.377040 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:45 crc kubenswrapper[4685]: E1204 06:26:45.377136 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:53.377104319 +0000 UTC m=+950.565335294 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "webhook-server-cert" not found Dec 04 06:26:45 crc kubenswrapper[4685]: E1204 06:26:45.377204 4685 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 06:26:45 crc kubenswrapper[4685]: E1204 06:26:45.377256 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:26:53.377243384 +0000 UTC m=+950.565474159 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "metrics-server-cert" not found Dec 04 06:26:52 crc kubenswrapper[4685]: I1204 06:26:52.196827 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:52 crc kubenswrapper[4685]: I1204 06:26:52.207035 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d55943b8-039d-4747-b341-d0fe2e0f6df7-cert\") pod \"infra-operator-controller-manager-57548d458d-8fjlw\" (UID: \"d55943b8-039d-4747-b341-d0fe2e0f6df7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:52 crc kubenswrapper[4685]: I1204 06:26:52.483527 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:26:52 crc kubenswrapper[4685]: I1204 06:26:52.706598 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:52 crc kubenswrapper[4685]: I1204 06:26:52.713617 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed35c5e-42ad-433c-a044-0fdac4c8ad06-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9\" (UID: \"3ed35c5e-42ad-433c-a044-0fdac4c8ad06\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:53 crc kubenswrapper[4685]: I1204 06:26:53.008817 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:26:53 crc kubenswrapper[4685]: I1204 06:26:53.417945 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:53 crc kubenswrapper[4685]: I1204 06:26:53.418115 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:53 crc kubenswrapper[4685]: E1204 06:26:53.418448 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 06:26:53 crc kubenswrapper[4685]: E1204 06:26:53.418544 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs podName:541675b0-83d5-49cf-8209-2de2d8385f3b nodeName:}" failed. No retries permitted until 2025-12-04 06:27:09.418524034 +0000 UTC m=+966.606754809 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs") pod "openstack-operator-controller-manager-54986f46d6-f8k5j" (UID: "541675b0-83d5-49cf-8209-2de2d8385f3b") : secret "webhook-server-cert" not found Dec 04 06:26:53 crc kubenswrapper[4685]: I1204 06:26:53.423962 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-metrics-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:26:57 crc kubenswrapper[4685]: E1204 06:26:57.224005 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 04 06:26:57 crc kubenswrapper[4685]: E1204 06:26:57.224856 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-948tx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-v2lxd_openstack-operators(09248202-ba20-4b70-95ab-1a7e67c04ef7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:26:57 crc kubenswrapper[4685]: E1204 06:26:57.844601 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 04 06:26:57 crc kubenswrapper[4685]: E1204 06:26:57.845425 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cbkq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-xchfb_openstack-operators(18233d1f-7861-4bd5-8643-f464fce185ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:26:58 crc kubenswrapper[4685]: E1204 06:26:58.405106 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 04 06:26:58 crc kubenswrapper[4685]: E1204 06:26:58.405286 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2p5wb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-rt2mx_openstack-operators(87b87b6f-538b-4ba3-ba4c-994924ffe76e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:26:58 crc kubenswrapper[4685]: E1204 06:26:58.908149 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 04 06:26:58 crc kubenswrapper[4685]: E1204 06:26:58.908461 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2zghz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-rrk46_openstack-operators(4af99755-81d3-43ec-a924-a3dcc4347315): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:26:59 crc kubenswrapper[4685]: E1204 06:26:59.464543 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 04 06:26:59 crc kubenswrapper[4685]: E1204 06:26:59.464751 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v2mbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-c64mf_openstack-operators(9cee5367-7803-44a8-b047-66e4311f9e0a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:26:59 crc kubenswrapper[4685]: E1204 06:26:59.986695 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 04 06:26:59 crc kubenswrapper[4685]: E1204 06:26:59.987067 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m4wl7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-h8fbl_openstack-operators(c9f1a1e0-3fd0-4c69-8475-9ed056330d57): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:27:00 crc kubenswrapper[4685]: E1204 06:27:00.467270 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 04 06:27:00 crc kubenswrapper[4685]: E1204 06:27:00.467768 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-97zmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-dnkjj_openstack-operators(89fc2488-fa6d-4515-923e-115aa45e23d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:27:00 crc kubenswrapper[4685]: E1204 06:27:00.882377 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 04 06:27:00 crc kubenswrapper[4685]: E1204 06:27:00.882655 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qnnv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-cdbkv_openstack-operators(efab633f-9698-443f-8175-279a82810d4d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:27:03 crc kubenswrapper[4685]: E1204 06:27:03.489105 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 04 06:27:03 crc kubenswrapper[4685]: E1204 06:27:03.489438 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qvb64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-xql6q_openstack-operators(cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:27:04 crc kubenswrapper[4685]: I1204 06:27:04.473661 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9"] Dec 04 06:27:04 crc kubenswrapper[4685]: I1204 06:27:04.560273 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw"] Dec 04 06:27:04 crc kubenswrapper[4685]: W1204 06:27:04.684280 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ed35c5e_42ad_433c_a044_0fdac4c8ad06.slice/crio-94550cc45e6faf8df0094e2bd807baa367ef554e97d93f8f43cfbae857bfa6d8 WatchSource:0}: Error finding container 94550cc45e6faf8df0094e2bd807baa367ef554e97d93f8f43cfbae857bfa6d8: Status 404 returned error can't find the container with id 94550cc45e6faf8df0094e2bd807baa367ef554e97d93f8f43cfbae857bfa6d8 Dec 04 06:27:04 crc kubenswrapper[4685]: W1204 06:27:04.686118 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd55943b8_039d_4747_b341_d0fe2e0f6df7.slice/crio-1e1acd863c0afb473c3c1628b2f7dc870c8d23a184b2efb555f538cacf632bb4 WatchSource:0}: Error finding container 1e1acd863c0afb473c3c1628b2f7dc870c8d23a184b2efb555f538cacf632bb4: Status 404 returned error can't find the container with id 1e1acd863c0afb473c3c1628b2f7dc870c8d23a184b2efb555f538cacf632bb4 Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.165132 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" event={"ID":"3ed35c5e-42ad-433c-a044-0fdac4c8ad06","Type":"ContainerStarted","Data":"94550cc45e6faf8df0094e2bd807baa367ef554e97d93f8f43cfbae857bfa6d8"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.168535 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" event={"ID":"c7545a63-4651-4d38-a251-9de38ce5a226","Type":"ContainerStarted","Data":"1cd9c8fd82daa402051a85eacba27b013a17f70bed9b40f451c50f164b76256a"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.169727 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" event={"ID":"e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13","Type":"ContainerStarted","Data":"e4a0c2c186c5169098bd471aa8cdb813219c7737d73a8707deea32ceec9c1cef"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.198356 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" event={"ID":"7aa55d54-0ab9-4b25-8226-095d3703ca03","Type":"ContainerStarted","Data":"0578feb9d977d42d0c04d3f6ce497652450e7d0016883b0f47e5e332a4e91d9d"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.220510 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" event={"ID":"211501da-f7e5-405e-b1cb-2673ca2b82c1","Type":"ContainerStarted","Data":"d4bb5b0c05716792814d47790c9385eb50ef95506596d8fb5af0e88f828524f9"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.224791 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" event={"ID":"231797d5-080c-4e7f-ad27-186a01bf4094","Type":"ContainerStarted","Data":"3af00e46ea564e74bfb4b59b5b574811e1502101318f235924a5addfd6fabf94"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.227067 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" event={"ID":"d55943b8-039d-4747-b341-d0fe2e0f6df7","Type":"ContainerStarted","Data":"1e1acd863c0afb473c3c1628b2f7dc870c8d23a184b2efb555f538cacf632bb4"} Dec 04 06:27:05 crc kubenswrapper[4685]: I1204 06:27:05.232919 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" event={"ID":"639879a9-dc01-4145-916b-8d568130af06","Type":"ContainerStarted","Data":"992e994d9a047f4e4dfc0e6c55da1e9fbac68e788cbe94f482bea0002661b973"} Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.121850 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.122308 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.122356 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.123130 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.123182 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11" gracePeriod=600 Dec 04 06:27:06 crc kubenswrapper[4685]: E1204 06:27:06.152117 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod416d3991_f0c3_40a2_b997_e40e43b4ddde.slice/crio-1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.244152 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" event={"ID":"4c3ba289-3038-4c41-881a-d53e4f54bc47","Type":"ContainerStarted","Data":"cf70b4298de067d13670dbe8f19668bdff39499ad66bb9c03e15bf536fd2528b"} Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.246022 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" event={"ID":"e9b56e59-1bc8-4608-bdfe-2901e30c7009","Type":"ContainerStarted","Data":"49d6da52da6f257b764eb86689f76ced169e0140e6ab851d4e2859d6debea809"} Dec 04 06:27:06 crc kubenswrapper[4685]: I1204 06:27:06.249112 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" event={"ID":"ee4e7fd4-46c8-4aac-be76-5097d9f76ac2","Type":"ContainerStarted","Data":"5c78dd9df09eebc9ab5aa247bd375f12a89f3e9d04855d89fc8e14758290857f"} Dec 04 06:27:06 crc kubenswrapper[4685]: E1204 06:27:06.897853 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" podUID="cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1" Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.261218 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" event={"ID":"639879a9-dc01-4145-916b-8d568130af06","Type":"ContainerStarted","Data":"7b1d55786e8a715d4236019da92a08a85a7c21e7fb4177e192ca7ca0e1d54f3a"} Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.261761 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:27:07 crc kubenswrapper[4685]: E1204 06:27:07.263631 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" podUID="87b87b6f-538b-4ba3-ba4c-994924ffe76e" Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.265448 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" event={"ID":"cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1","Type":"ContainerStarted","Data":"0a96caae962bb2a51222e221b66b425778177a177dd9c0b266cea4dca4d2049d"} Dec 04 06:27:07 crc kubenswrapper[4685]: E1204 06:27:07.271272 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" podUID="cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1" Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.274488 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" event={"ID":"e3ac8015-c6a6-4837-9118-0de5f1b351df","Type":"ContainerStarted","Data":"485f19be913d718fac93076a0f6ab48c2909fb6bccdd3c6d8c9f76c2198db7fd"} Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.302799 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" event={"ID":"ecb2e435-29cf-4b37-a0bb-4f2441cab009","Type":"ContainerStarted","Data":"b4893b952424c22972a16bce85c7d6ee9d6aa52bc7e613e998cea57ef6edc811"} Dec 04 06:27:07 crc kubenswrapper[4685]: E1204 06:27:07.310634 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" podUID="9cee5367-7803-44a8-b047-66e4311f9e0a" Dec 04 06:27:07 crc kubenswrapper[4685]: E1204 06:27:07.310755 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" podUID="87b87b6f-538b-4ba3-ba4c-994924ffe76e" Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.323604 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" podStartSLOduration=3.15347042 podStartE2EDuration="31.323575124s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.712739154 +0000 UTC m=+935.900969929" lastFinishedPulling="2025-12-04 06:27:06.882843838 +0000 UTC m=+964.071074633" observedRunningTime="2025-12-04 06:27:07.290570043 +0000 UTC m=+964.478800818" watchObservedRunningTime="2025-12-04 06:27:07.323575124 +0000 UTC m=+964.511805899" Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.328324 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11" exitCode=0 Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.328380 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11"} Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.328478 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"bca40440ad9d82bdf8b013f3ae5e8f16199c54c667e88ee4d71345b125617d55"} Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.328509 4685 scope.go:117] "RemoveContainer" containerID="39d6b27b9ba56cb920b538ba51d04d90d1b790d65f04ecd9199b893b79b8e599" Dec 04 06:27:07 crc kubenswrapper[4685]: I1204 06:27:07.402958 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tmmg4" podStartSLOduration=4.489467525 podStartE2EDuration="30.402937788s" podCreationTimestamp="2025-12-04 06:26:37 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.79087403 +0000 UTC m=+935.979104795" lastFinishedPulling="2025-12-04 06:27:04.704344283 +0000 UTC m=+961.892575058" observedRunningTime="2025-12-04 06:27:07.388153639 +0000 UTC m=+964.576384414" watchObservedRunningTime="2025-12-04 06:27:07.402937788 +0000 UTC m=+964.591168563" Dec 04 06:27:07 crc kubenswrapper[4685]: E1204 06:27:07.455648 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" podUID="89fc2488-fa6d-4515-923e-115aa45e23d4" Dec 04 06:27:07 crc kubenswrapper[4685]: E1204 06:27:07.510752 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" podUID="c9f1a1e0-3fd0-4c69-8475-9ed056330d57" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.337828 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" event={"ID":"211501da-f7e5-405e-b1cb-2673ca2b82c1","Type":"ContainerStarted","Data":"35a11cda2f512b8b27cd124995d5d29c1079c219a8c55ab5aa5c077f46d38321"} Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.338453 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.342636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" event={"ID":"7aa55d54-0ab9-4b25-8226-095d3703ca03","Type":"ContainerStarted","Data":"57d393d4ec9c66086ad293644fc620768ad05cc384e35158c5ed69cfd6cfe614"} Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.342777 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.345069 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" event={"ID":"c9f1a1e0-3fd0-4c69-8475-9ed056330d57","Type":"ContainerStarted","Data":"69f5d7970d7803a3afba830e6511eb548cad6fddf8e693bf59c1320dc3ad19e7"} Dec 04 06:27:08 crc kubenswrapper[4685]: E1204 06:27:08.347215 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" podUID="c9f1a1e0-3fd0-4c69-8475-9ed056330d57" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.348041 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" event={"ID":"89fc2488-fa6d-4515-923e-115aa45e23d4","Type":"ContainerStarted","Data":"390356fa2ee046bfe6a51cff995a6af8accaa14f242294be44bd9fecb1910aba"} Dec 04 06:27:08 crc kubenswrapper[4685]: E1204 06:27:08.350015 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" podUID="89fc2488-fa6d-4515-923e-115aa45e23d4" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.359786 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" event={"ID":"e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13","Type":"ContainerStarted","Data":"e677dd0cb49457e31c51f30f6683fcd31da42649af0ac02c92f906cb19bb475a"} Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.360136 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" podStartSLOduration=3.381365292 podStartE2EDuration="32.360114008s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:37.904537106 +0000 UTC m=+935.092767881" lastFinishedPulling="2025-12-04 06:27:06.883285822 +0000 UTC m=+964.071516597" observedRunningTime="2025-12-04 06:27:08.358948621 +0000 UTC m=+965.547179396" watchObservedRunningTime="2025-12-04 06:27:08.360114008 +0000 UTC m=+965.548344783" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.361702 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.369452 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" event={"ID":"87b87b6f-538b-4ba3-ba4c-994924ffe76e","Type":"ContainerStarted","Data":"6c2e5450540c94bc51dcb92427e331b50e4e51b652dba0738c24f5195117ee64"} Dec 04 06:27:08 crc kubenswrapper[4685]: E1204 06:27:08.371932 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" podUID="87b87b6f-538b-4ba3-ba4c-994924ffe76e" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.373388 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" event={"ID":"9cee5367-7803-44a8-b047-66e4311f9e0a","Type":"ContainerStarted","Data":"baf85900ce4e71a1c9a252b8aa196cd1f54f758af19078f5241fd9b4e7eaa2ba"} Dec 04 06:27:08 crc kubenswrapper[4685]: E1204 06:27:08.377503 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" podUID="9cee5367-7803-44a8-b047-66e4311f9e0a" Dec 04 06:27:08 crc kubenswrapper[4685]: E1204 06:27:08.377968 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" podUID="cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.394849 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" podStartSLOduration=3.879043429 podStartE2EDuration="32.394821503s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.347471321 +0000 UTC m=+935.535702096" lastFinishedPulling="2025-12-04 06:27:06.863249405 +0000 UTC m=+964.051480170" observedRunningTime="2025-12-04 06:27:08.385748184 +0000 UTC m=+965.573978959" watchObservedRunningTime="2025-12-04 06:27:08.394821503 +0000 UTC m=+965.583052278" Dec 04 06:27:08 crc kubenswrapper[4685]: I1204 06:27:08.486530 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" podStartSLOduration=4.428802143 podStartE2EDuration="32.48649725s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.796308613 +0000 UTC m=+935.984539388" lastFinishedPulling="2025-12-04 06:27:06.85400372 +0000 UTC m=+964.042234495" observedRunningTime="2025-12-04 06:27:08.482190003 +0000 UTC m=+965.670420798" watchObservedRunningTime="2025-12-04 06:27:08.48649725 +0000 UTC m=+965.674728045" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.384732 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" event={"ID":"e9b56e59-1bc8-4608-bdfe-2901e30c7009","Type":"ContainerStarted","Data":"1cbd72af82b807a39c01d1fe3420b5098df77a755408b4e33113b7c505872190"} Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.385521 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.389739 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" event={"ID":"e3ac8015-c6a6-4837-9118-0de5f1b351df","Type":"ContainerStarted","Data":"dddeaab1668a2970c7b480f8de3f5a3e4d196bbe761c36cac1431e222d16f50a"} Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.390557 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.392148 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" event={"ID":"ee4e7fd4-46c8-4aac-be76-5097d9f76ac2","Type":"ContainerStarted","Data":"428b11afe9127329a688763a47bf997afe6a2d3ccf8d20de81cfd3af15b6b69a"} Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.392628 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.394126 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" event={"ID":"231797d5-080c-4e7f-ad27-186a01bf4094","Type":"ContainerStarted","Data":"b357aec3bf4cd7252c26a71da0dca3e048e438f2f81a07e9ebf443b50294fd6d"} Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.394657 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.396918 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" event={"ID":"4c3ba289-3038-4c41-881a-d53e4f54bc47","Type":"ContainerStarted","Data":"882986a7fbda4568b19f88e13135d332f47e1ed5eaf9a18ff502092c0619893d"} Dec 04 06:27:09 crc kubenswrapper[4685]: E1204 06:27:09.400362 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" podUID="89fc2488-fa6d-4515-923e-115aa45e23d4" Dec 04 06:27:09 crc kubenswrapper[4685]: E1204 06:27:09.400825 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" podUID="9cee5367-7803-44a8-b047-66e4311f9e0a" Dec 04 06:27:09 crc kubenswrapper[4685]: E1204 06:27:09.401068 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" podUID="c9f1a1e0-3fd0-4c69-8475-9ed056330d57" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.412850 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-wd5jl" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.414078 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" podStartSLOduration=4.79229488 podStartE2EDuration="33.413890082s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.326648088 +0000 UTC m=+935.514878863" lastFinishedPulling="2025-12-04 06:27:06.94824329 +0000 UTC m=+964.136474065" observedRunningTime="2025-12-04 06:27:09.402652004 +0000 UTC m=+966.590882789" watchObservedRunningTime="2025-12-04 06:27:09.413890082 +0000 UTC m=+966.602120867" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.438899 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" podStartSLOduration=4.920421648 podStartE2EDuration="33.438875807s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.344694142 +0000 UTC m=+935.532924917" lastFinishedPulling="2025-12-04 06:27:06.863148301 +0000 UTC m=+964.051379076" observedRunningTime="2025-12-04 06:27:09.430917214 +0000 UTC m=+966.619147999" watchObservedRunningTime="2025-12-04 06:27:09.438875807 +0000 UTC m=+966.627106582" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.475749 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.482269 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" podStartSLOduration=4.164170342 podStartE2EDuration="33.482238097s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:37.623692348 +0000 UTC m=+934.811923123" lastFinishedPulling="2025-12-04 06:27:06.941760103 +0000 UTC m=+964.129990878" observedRunningTime="2025-12-04 06:27:09.464537193 +0000 UTC m=+966.652767968" watchObservedRunningTime="2025-12-04 06:27:09.482238097 +0000 UTC m=+966.670468872" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.484370 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/541675b0-83d5-49cf-8209-2de2d8385f3b-webhook-certs\") pod \"openstack-operator-controller-manager-54986f46d6-f8k5j\" (UID: \"541675b0-83d5-49cf-8209-2de2d8385f3b\") " pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.524116 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" podStartSLOduration=4.126888977 podStartE2EDuration="33.524094409s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:37.554477596 +0000 UTC m=+934.742708371" lastFinishedPulling="2025-12-04 06:27:06.951683028 +0000 UTC m=+964.139913803" observedRunningTime="2025-12-04 06:27:09.51562935 +0000 UTC m=+966.703860125" watchObservedRunningTime="2025-12-04 06:27:09.524094409 +0000 UTC m=+966.712325184" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.544738 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" podStartSLOduration=8.323881553 podStartE2EDuration="33.544718365s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.790819729 +0000 UTC m=+935.979050504" lastFinishedPulling="2025-12-04 06:27:04.011656531 +0000 UTC m=+961.199887316" observedRunningTime="2025-12-04 06:27:09.540632695 +0000 UTC m=+966.728863460" watchObservedRunningTime="2025-12-04 06:27:09.544718365 +0000 UTC m=+966.732949130" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.771879 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-f7csq" Dec 04 06:27:09 crc kubenswrapper[4685]: I1204 06:27:09.780529 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:27:10 crc kubenswrapper[4685]: E1204 06:27:10.012957 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" podUID="18233d1f-7861-4bd5-8643-f464fce185ca" Dec 04 06:27:10 crc kubenswrapper[4685]: E1204 06:27:10.059400 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" podUID="09248202-ba20-4b70-95ab-1a7e67c04ef7" Dec 04 06:27:10 crc kubenswrapper[4685]: E1204 06:27:10.078241 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" podUID="4af99755-81d3-43ec-a924-a3dcc4347315" Dec 04 06:27:10 crc kubenswrapper[4685]: E1204 06:27:10.090660 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" podUID="efab633f-9698-443f-8175-279a82810d4d" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.257783 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j"] Dec 04 06:27:10 crc kubenswrapper[4685]: W1204 06:27:10.259699 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod541675b0_83d5_49cf_8209_2de2d8385f3b.slice/crio-8d8885849c678614d9c2a18d695452888ded23e04753be7fa9326f1f1a296e69 WatchSource:0}: Error finding container 8d8885849c678614d9c2a18d695452888ded23e04753be7fa9326f1f1a296e69: Status 404 returned error can't find the container with id 8d8885849c678614d9c2a18d695452888ded23e04753be7fa9326f1f1a296e69 Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.406889 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" event={"ID":"541675b0-83d5-49cf-8209-2de2d8385f3b","Type":"ContainerStarted","Data":"8d8885849c678614d9c2a18d695452888ded23e04753be7fa9326f1f1a296e69"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.410891 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" event={"ID":"d55943b8-039d-4747-b341-d0fe2e0f6df7","Type":"ContainerStarted","Data":"e8d14176e4e78639aaf3c1406fd5419382f20552efd22b714d4dfc35f06b4143"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.410922 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" event={"ID":"d55943b8-039d-4747-b341-d0fe2e0f6df7","Type":"ContainerStarted","Data":"212ad372694aeaa3d89304973da143da4e0163e2c40d1b52dd8bb34eaed87d28"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.412288 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.415053 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" event={"ID":"3ed35c5e-42ad-433c-a044-0fdac4c8ad06","Type":"ContainerStarted","Data":"9724a32b50cc223514da4579256d34867f86abbddf53b0f6c1a82a85e2873a25"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.415084 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" event={"ID":"3ed35c5e-42ad-433c-a044-0fdac4c8ad06","Type":"ContainerStarted","Data":"159b72c61b40108c793eeacdb6da8d9dbbea47d984b8a01cc1bf85284a30b053"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.415550 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.417734 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" event={"ID":"4af99755-81d3-43ec-a924-a3dcc4347315","Type":"ContainerStarted","Data":"63eb3366642c2ce6ad7c9a05f9956ed91442e75eb98cf730195c2d364612a1d7"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.421395 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" event={"ID":"efab633f-9698-443f-8175-279a82810d4d","Type":"ContainerStarted","Data":"39b6f6c596c9e1315856787b0402f629d2bfe7dc4005ba9f2d28add17e48ab72"} Dec 04 06:27:10 crc kubenswrapper[4685]: E1204 06:27:10.422522 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" podUID="efab633f-9698-443f-8175-279a82810d4d" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.433831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" event={"ID":"18233d1f-7861-4bd5-8643-f464fce185ca","Type":"ContainerStarted","Data":"f3120d953a2a60f64d3925d73820b12d4e90f52eb900fa66eb2894d2f4492032"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.439764 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" podStartSLOduration=29.431556154 podStartE2EDuration="34.439749757s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:27:04.689459919 +0000 UTC m=+961.877690694" lastFinishedPulling="2025-12-04 06:27:09.697653522 +0000 UTC m=+966.885884297" observedRunningTime="2025-12-04 06:27:10.431292108 +0000 UTC m=+967.619522883" watchObservedRunningTime="2025-12-04 06:27:10.439749757 +0000 UTC m=+967.627980522" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.442356 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" event={"ID":"c7545a63-4651-4d38-a251-9de38ce5a226","Type":"ContainerStarted","Data":"13cbfd8ed5d1a71ded8ce205f68756877cf6e23274e54c102331cec78adefbca"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.443601 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.448223 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" event={"ID":"09248202-ba20-4b70-95ab-1a7e67c04ef7","Type":"ContainerStarted","Data":"0dcf0fbbc0eeb0b5162c3f01132e71110691312195cff454bd822a87ec7d5499"} Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.451719 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.451778 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7t9xh" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.452300 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.454335 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qqd9p" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.454513 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-czngd" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.454633 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-s8zzd" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.542168 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" podStartSLOduration=29.574898847 podStartE2EDuration="34.542130315s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:27:04.686167575 +0000 UTC m=+961.874398350" lastFinishedPulling="2025-12-04 06:27:09.653399043 +0000 UTC m=+966.841629818" observedRunningTime="2025-12-04 06:27:10.513491013 +0000 UTC m=+967.701721788" watchObservedRunningTime="2025-12-04 06:27:10.542130315 +0000 UTC m=+967.730361090" Dec 04 06:27:10 crc kubenswrapper[4685]: I1204 06:27:10.596556 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jlm87" podStartSLOduration=3.606182516 podStartE2EDuration="34.596532516s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.708773898 +0000 UTC m=+935.897004673" lastFinishedPulling="2025-12-04 06:27:09.699123898 +0000 UTC m=+966.887354673" observedRunningTime="2025-12-04 06:27:10.589200022 +0000 UTC m=+967.777430797" watchObservedRunningTime="2025-12-04 06:27:10.596532516 +0000 UTC m=+967.784763291" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.456507 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" event={"ID":"09248202-ba20-4b70-95ab-1a7e67c04ef7","Type":"ContainerStarted","Data":"db61577de69bc0fe33498d573becec19bbfb3bb209a395a9ae41954ea004bc03"} Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.457063 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.459617 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" event={"ID":"4af99755-81d3-43ec-a924-a3dcc4347315","Type":"ContainerStarted","Data":"0c72aac765f56f843c5ee7a8b914d1e2ae4c037cd373ea79a70011c7e7ed548e"} Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.459777 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.461716 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" event={"ID":"18233d1f-7861-4bd5-8643-f464fce185ca","Type":"ContainerStarted","Data":"0f9be1707ffba6a48954d1b859655427d27545d8e7e0860eef980c7167136cc7"} Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.462010 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.463197 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" event={"ID":"541675b0-83d5-49cf-8209-2de2d8385f3b","Type":"ContainerStarted","Data":"2c46a73d97d7531a781f711471bf5ab8ca16f1d17aeefbddddada06b070bef30"} Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.464261 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.466766 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7wr8" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.508188 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" podStartSLOduration=3.377393516 podStartE2EDuration="35.508160456s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.763910633 +0000 UTC m=+935.952141408" lastFinishedPulling="2025-12-04 06:27:10.894677573 +0000 UTC m=+968.082908348" observedRunningTime="2025-12-04 06:27:11.503933241 +0000 UTC m=+968.692164046" watchObservedRunningTime="2025-12-04 06:27:11.508160456 +0000 UTC m=+968.696391251" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.533665 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" podStartSLOduration=3.414498636 podStartE2EDuration="35.533646637s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.77453177 +0000 UTC m=+935.962762545" lastFinishedPulling="2025-12-04 06:27:10.893679771 +0000 UTC m=+968.081910546" observedRunningTime="2025-12-04 06:27:11.532012444 +0000 UTC m=+968.720243219" watchObservedRunningTime="2025-12-04 06:27:11.533646637 +0000 UTC m=+968.721877412" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.559521 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" podStartSLOduration=35.55949265 podStartE2EDuration="35.55949265s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:27:11.555660527 +0000 UTC m=+968.743891302" watchObservedRunningTime="2025-12-04 06:27:11.55949265 +0000 UTC m=+968.747723425" Dec 04 06:27:11 crc kubenswrapper[4685]: I1204 06:27:11.598045 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" podStartSLOduration=3.491266578 podStartE2EDuration="35.598015515s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.790607012 +0000 UTC m=+935.978837787" lastFinishedPulling="2025-12-04 06:27:10.897355939 +0000 UTC m=+968.085586724" observedRunningTime="2025-12-04 06:27:11.597431837 +0000 UTC m=+968.785662602" watchObservedRunningTime="2025-12-04 06:27:11.598015515 +0000 UTC m=+968.786246290" Dec 04 06:27:12 crc kubenswrapper[4685]: I1204 06:27:12.474929 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" event={"ID":"efab633f-9698-443f-8175-279a82810d4d","Type":"ContainerStarted","Data":"920db66d1ea4c34c2d2daff1a00390fe82be7cc40319c0725ea962701b0525d9"} Dec 04 06:27:12 crc kubenswrapper[4685]: I1204 06:27:12.510620 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" podStartSLOduration=3.096699792 podStartE2EDuration="36.510579225s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.713278181 +0000 UTC m=+935.901508956" lastFinishedPulling="2025-12-04 06:27:12.127157624 +0000 UTC m=+969.315388389" observedRunningTime="2025-12-04 06:27:12.509529961 +0000 UTC m=+969.697760776" watchObservedRunningTime="2025-12-04 06:27:12.510579225 +0000 UTC m=+969.698810010" Dec 04 06:27:16 crc kubenswrapper[4685]: I1204 06:27:16.460395 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6pgsq" Dec 04 06:27:16 crc kubenswrapper[4685]: I1204 06:27:16.812065 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rrk46" Dec 04 06:27:17 crc kubenswrapper[4685]: I1204 06:27:17.037258 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-v2lxd" Dec 04 06:27:17 crc kubenswrapper[4685]: I1204 06:27:17.087644 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:27:17 crc kubenswrapper[4685]: I1204 06:27:17.088777 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cdbkv" Dec 04 06:27:17 crc kubenswrapper[4685]: I1204 06:27:17.267659 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xchfb" Dec 04 06:27:17 crc kubenswrapper[4685]: I1204 06:27:17.374806 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nhb96" Dec 04 06:27:17 crc kubenswrapper[4685]: I1204 06:27:17.753622 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-zzn46" Dec 04 06:27:19 crc kubenswrapper[4685]: I1204 06:27:19.787243 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-54986f46d6-f8k5j" Dec 04 06:27:20 crc kubenswrapper[4685]: I1204 06:27:20.556812 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" event={"ID":"cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1","Type":"ContainerStarted","Data":"e5c7a4da0a91bff3bb23976d046e3091fdc9ba859680f9cd342d0d116ad71a77"} Dec 04 06:27:20 crc kubenswrapper[4685]: I1204 06:27:20.558965 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:27:20 crc kubenswrapper[4685]: I1204 06:27:20.583972 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" podStartSLOduration=3.029604108 podStartE2EDuration="44.583945077s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.344700343 +0000 UTC m=+935.532931118" lastFinishedPulling="2025-12-04 06:27:19.899041322 +0000 UTC m=+977.087272087" observedRunningTime="2025-12-04 06:27:20.580316309 +0000 UTC m=+977.768547104" watchObservedRunningTime="2025-12-04 06:27:20.583945077 +0000 UTC m=+977.772175852" Dec 04 06:27:21 crc kubenswrapper[4685]: I1204 06:27:21.577506 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" event={"ID":"87b87b6f-538b-4ba3-ba4c-994924ffe76e","Type":"ContainerStarted","Data":"b9b436e7681dcda14aeb536c0c411788a8c9502fef7542daa75d262114308416"} Dec 04 06:27:21 crc kubenswrapper[4685]: I1204 06:27:21.577795 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:27:21 crc kubenswrapper[4685]: I1204 06:27:21.603698 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" podStartSLOduration=2.853827 podStartE2EDuration="45.603675524s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:37.828870408 +0000 UTC m=+935.017101183" lastFinishedPulling="2025-12-04 06:27:20.578718932 +0000 UTC m=+977.766949707" observedRunningTime="2025-12-04 06:27:21.598699146 +0000 UTC m=+978.786929951" watchObservedRunningTime="2025-12-04 06:27:21.603675524 +0000 UTC m=+978.791906309" Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.501009 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8fjlw" Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.585054 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" event={"ID":"89fc2488-fa6d-4515-923e-115aa45e23d4","Type":"ContainerStarted","Data":"65327d00aef2ac879359c764cbad75dc818fe7b62e5c54cfbbf54313085d791d"} Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.585344 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.587302 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" event={"ID":"9cee5367-7803-44a8-b047-66e4311f9e0a","Type":"ContainerStarted","Data":"fa3ec262c6871ee7427d4c1bc027619ebe4f9a67a2f16256e19a628f48c62187"} Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.587670 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.604474 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" podStartSLOduration=3.438868205 podStartE2EDuration="46.604451457s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.359870895 +0000 UTC m=+935.548101670" lastFinishedPulling="2025-12-04 06:27:21.525454147 +0000 UTC m=+978.713684922" observedRunningTime="2025-12-04 06:27:22.602389826 +0000 UTC m=+979.790620631" watchObservedRunningTime="2025-12-04 06:27:22.604451457 +0000 UTC m=+979.792682232" Dec 04 06:27:22 crc kubenswrapper[4685]: I1204 06:27:22.619386 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" podStartSLOduration=3.826662369 podStartE2EDuration="46.61935111s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:38.763787089 +0000 UTC m=+935.952017864" lastFinishedPulling="2025-12-04 06:27:21.55647581 +0000 UTC m=+978.744706605" observedRunningTime="2025-12-04 06:27:22.617348751 +0000 UTC m=+979.805579526" watchObservedRunningTime="2025-12-04 06:27:22.61935111 +0000 UTC m=+979.807581885" Dec 04 06:27:23 crc kubenswrapper[4685]: I1204 06:27:23.018248 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9" Dec 04 06:27:26 crc kubenswrapper[4685]: I1204 06:27:26.498174 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rt2mx" Dec 04 06:27:26 crc kubenswrapper[4685]: I1204 06:27:26.736594 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-xql6q" Dec 04 06:27:26 crc kubenswrapper[4685]: I1204 06:27:26.846105 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-dnkjj" Dec 04 06:27:27 crc kubenswrapper[4685]: I1204 06:27:27.345690 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-c64mf" Dec 04 06:27:33 crc kubenswrapper[4685]: I1204 06:27:33.716551 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" event={"ID":"c9f1a1e0-3fd0-4c69-8475-9ed056330d57","Type":"ContainerStarted","Data":"0551407ad1b421780e1ae4534b7d6152a9ff1f9a8cd69478581d9fac3107231e"} Dec 04 06:27:33 crc kubenswrapper[4685]: I1204 06:27:33.717980 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:27:33 crc kubenswrapper[4685]: I1204 06:27:33.743880 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" podStartSLOduration=2.224767511 podStartE2EDuration="57.743839964s" podCreationTimestamp="2025-12-04 06:26:36 +0000 UTC" firstStartedPulling="2025-12-04 06:26:37.300531915 +0000 UTC m=+934.488762690" lastFinishedPulling="2025-12-04 06:27:32.819604358 +0000 UTC m=+990.007835143" observedRunningTime="2025-12-04 06:27:33.736894548 +0000 UTC m=+990.925125343" watchObservedRunningTime="2025-12-04 06:27:33.743839964 +0000 UTC m=+990.932070749" Dec 04 06:27:46 crc kubenswrapper[4685]: I1204 06:27:46.369396 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h8fbl" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.413138 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lhd8w"] Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.415024 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.416960 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.417384 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.417772 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.419661 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-vj8ct" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.433883 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lhd8w"] Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.451570 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ppmsn"] Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.453886 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.455517 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.479463 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ppmsn"] Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.503090 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvwkx\" (UniqueName: \"kubernetes.io/projected/db42e3fb-f38f-4892-bf85-de67095bb931-kube-api-access-xvwkx\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.503136 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be665a16-e393-4fe7-9f6b-c6412d555eb1-config\") pod \"dnsmasq-dns-675f4bcbfc-lhd8w\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.503161 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szvmn\" (UniqueName: \"kubernetes.io/projected/be665a16-e393-4fe7-9f6b-c6412d555eb1-kube-api-access-szvmn\") pod \"dnsmasq-dns-675f4bcbfc-lhd8w\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.503190 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-config\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.503493 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.604880 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.604976 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvwkx\" (UniqueName: \"kubernetes.io/projected/db42e3fb-f38f-4892-bf85-de67095bb931-kube-api-access-xvwkx\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.605014 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be665a16-e393-4fe7-9f6b-c6412d555eb1-config\") pod \"dnsmasq-dns-675f4bcbfc-lhd8w\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.605035 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szvmn\" (UniqueName: \"kubernetes.io/projected/be665a16-e393-4fe7-9f6b-c6412d555eb1-kube-api-access-szvmn\") pod \"dnsmasq-dns-675f4bcbfc-lhd8w\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.605064 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-config\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.606220 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.606231 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be665a16-e393-4fe7-9f6b-c6412d555eb1-config\") pod \"dnsmasq-dns-675f4bcbfc-lhd8w\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.606236 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-config\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.629458 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szvmn\" (UniqueName: \"kubernetes.io/projected/be665a16-e393-4fe7-9f6b-c6412d555eb1-kube-api-access-szvmn\") pod \"dnsmasq-dns-675f4bcbfc-lhd8w\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.638686 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvwkx\" (UniqueName: \"kubernetes.io/projected/db42e3fb-f38f-4892-bf85-de67095bb931-kube-api-access-xvwkx\") pod \"dnsmasq-dns-78dd6ddcc-ppmsn\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.734798 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:03 crc kubenswrapper[4685]: I1204 06:28:03.773819 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:04 crc kubenswrapper[4685]: I1204 06:28:04.069550 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lhd8w"] Dec 04 06:28:04 crc kubenswrapper[4685]: I1204 06:28:04.077511 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:28:04 crc kubenswrapper[4685]: I1204 06:28:04.319382 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ppmsn"] Dec 04 06:28:05 crc kubenswrapper[4685]: I1204 06:28:05.006527 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" event={"ID":"be665a16-e393-4fe7-9f6b-c6412d555eb1","Type":"ContainerStarted","Data":"10d0e5af7121121d0f8533e1e7dca45c2a9ebc540f9073b3f6a6fedaa6375e24"} Dec 04 06:28:05 crc kubenswrapper[4685]: I1204 06:28:05.009143 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" event={"ID":"db42e3fb-f38f-4892-bf85-de67095bb931","Type":"ContainerStarted","Data":"8df0a712864f5e78b675cf41ae2dff4482e9bf416d0544b2b81f0a7571c109e5"} Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.337445 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lhd8w"] Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.366244 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p5754"] Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.368959 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.376774 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p5754"] Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.456700 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.456801 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlzt9\" (UniqueName: \"kubernetes.io/projected/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-kube-api-access-zlzt9\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.456860 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-config\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.560530 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-config\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.560656 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.560695 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlzt9\" (UniqueName: \"kubernetes.io/projected/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-kube-api-access-zlzt9\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.562527 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.563837 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-config\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.585904 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlzt9\" (UniqueName: \"kubernetes.io/projected/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-kube-api-access-zlzt9\") pod \"dnsmasq-dns-666b6646f7-p5754\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.636224 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ppmsn"] Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.668358 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6wc8c"] Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.669615 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.686221 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6wc8c"] Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.701675 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.765186 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-config\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.765811 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lkcc\" (UniqueName: \"kubernetes.io/projected/6f031df1-b162-4ace-8c47-021a7d1c9e61-kube-api-access-5lkcc\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.765853 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.868285 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lkcc\" (UniqueName: \"kubernetes.io/projected/6f031df1-b162-4ace-8c47-021a7d1c9e61-kube-api-access-5lkcc\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.868362 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.868457 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-config\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.869722 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-config\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.870452 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:06 crc kubenswrapper[4685]: I1204 06:28:06.887552 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lkcc\" (UniqueName: \"kubernetes.io/projected/6f031df1-b162-4ace-8c47-021a7d1c9e61-kube-api-access-5lkcc\") pod \"dnsmasq-dns-57d769cc4f-6wc8c\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.017217 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.162801 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p5754"] Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.511566 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.534158 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.534307 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.541934 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9qfd7" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.542260 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.542559 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.542564 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.542657 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.543340 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.543384 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.594178 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/777e7995-c510-4d30-bcf1-d479e7cfa432-pod-info\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.594234 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.595350 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.595716 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ngrc\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-kube-api-access-4ngrc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.595794 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.595870 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-server-conf\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.595965 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-config-data\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.596029 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.596087 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.596117 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/777e7995-c510-4d30-bcf1-d479e7cfa432-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.596171 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698080 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ngrc\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-kube-api-access-4ngrc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698172 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698223 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-server-conf\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698257 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-config-data\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698288 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698319 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698343 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/777e7995-c510-4d30-bcf1-d479e7cfa432-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698398 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/777e7995-c510-4d30-bcf1-d479e7cfa432-pod-info\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698445 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698475 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698651 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.698972 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.699354 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-config-data\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.700562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.700961 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-server-conf\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.702213 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.706083 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/777e7995-c510-4d30-bcf1-d479e7cfa432-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.706094 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/777e7995-c510-4d30-bcf1-d479e7cfa432-pod-info\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.706735 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.707885 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.717467 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ngrc\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-kube-api-access-4ngrc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.745526 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.831352 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.832640 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.835245 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.835341 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.835503 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.851775 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.835690 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-nlrml" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.835683 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.841298 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.842976 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.873530 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904280 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw244\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-kube-api-access-sw244\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904309 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904353 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904386 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ff4ac6f-e715-439b-8b4a-2b2715198dab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904447 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904476 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904519 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904613 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ff4ac6f-e715-439b-8b4a-2b2715198dab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:07 crc kubenswrapper[4685]: I1204 06:28:07.904679 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.006698 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.007500 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.007538 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.007572 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008181 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008427 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ff4ac6f-e715-439b-8b4a-2b2715198dab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008491 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008616 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008840 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw244\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-kube-api-access-sw244\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008888 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008912 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008929 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.008946 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ff4ac6f-e715-439b-8b4a-2b2715198dab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.009543 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.010057 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.010080 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.012035 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.016590 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.017582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ff4ac6f-e715-439b-8b4a-2b2715198dab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.021001 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ff4ac6f-e715-439b-8b4a-2b2715198dab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.023970 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.031769 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw244\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-kube-api-access-sw244\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.067045 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:08 crc kubenswrapper[4685]: I1204 06:28:08.207900 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.371741 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.373808 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.379470 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.381489 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.381612 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.381884 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.382714 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-z6bmz" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.394932 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541459 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-operator-scripts\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541523 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577f79bb-8577-412a-92e1-bf56244becce-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541551 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-config-data-default\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541592 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541643 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/577f79bb-8577-412a-92e1-bf56244becce-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541665 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55rb4\" (UniqueName: \"kubernetes.io/projected/577f79bb-8577-412a-92e1-bf56244becce-kube-api-access-55rb4\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541687 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-kolla-config\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.541705 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/577f79bb-8577-412a-92e1-bf56244becce-config-data-generated\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643623 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/577f79bb-8577-412a-92e1-bf56244becce-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643690 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55rb4\" (UniqueName: \"kubernetes.io/projected/577f79bb-8577-412a-92e1-bf56244becce-kube-api-access-55rb4\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643719 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-kolla-config\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643745 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/577f79bb-8577-412a-92e1-bf56244becce-config-data-generated\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643848 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-operator-scripts\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643882 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577f79bb-8577-412a-92e1-bf56244becce-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643911 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-config-data-default\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.645017 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.643950 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.646041 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-operator-scripts\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.647462 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-kolla-config\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.648553 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/577f79bb-8577-412a-92e1-bf56244becce-config-data-default\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.650232 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/577f79bb-8577-412a-92e1-bf56244becce-config-data-generated\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.659146 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577f79bb-8577-412a-92e1-bf56244becce-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.683909 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55rb4\" (UniqueName: \"kubernetes.io/projected/577f79bb-8577-412a-92e1-bf56244becce-kube-api-access-55rb4\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.685043 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/577f79bb-8577-412a-92e1-bf56244becce-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:09 crc kubenswrapper[4685]: I1204 06:28:09.744491 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"577f79bb-8577-412a-92e1-bf56244becce\") " pod="openstack/openstack-galera-0" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.009850 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.911261 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.913573 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.916592 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-57f7r" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.916913 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.917599 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.918505 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 06:28:10 crc kubenswrapper[4685]: I1204 06:28:10.948340 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.075665 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.075757 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.076034 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.076122 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.076157 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmgv6\" (UniqueName: \"kubernetes.io/projected/536ebc06-2d82-419d-ac24-957fbc6734c9-kube-api-access-lmgv6\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.076361 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/536ebc06-2d82-419d-ac24-957fbc6734c9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.076494 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536ebc06-2d82-419d-ac24-957fbc6734c9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.076544 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/536ebc06-2d82-419d-ac24-957fbc6734c9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.178741 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/536ebc06-2d82-419d-ac24-957fbc6734c9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.178820 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536ebc06-2d82-419d-ac24-957fbc6734c9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.178843 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/536ebc06-2d82-419d-ac24-957fbc6734c9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.178909 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.178940 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.178985 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.179007 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.179031 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmgv6\" (UniqueName: \"kubernetes.io/projected/536ebc06-2d82-419d-ac24-957fbc6734c9-kube-api-access-lmgv6\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.179859 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.181031 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.181574 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.181936 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/536ebc06-2d82-419d-ac24-957fbc6734c9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.182848 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536ebc06-2d82-419d-ac24-957fbc6734c9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.187685 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/536ebc06-2d82-419d-ac24-957fbc6734c9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.188316 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536ebc06-2d82-419d-ac24-957fbc6734c9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.207055 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.209034 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.212917 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-b8sbf" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.213126 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.213467 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.212935 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmgv6\" (UniqueName: \"kubernetes.io/projected/536ebc06-2d82-419d-ac24-957fbc6734c9-kube-api-access-lmgv6\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.219540 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"536ebc06-2d82-419d-ac24-957fbc6734c9\") " pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.247603 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.249423 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.281252 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgcbk\" (UniqueName: \"kubernetes.io/projected/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-kube-api-access-mgcbk\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.281389 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.281463 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.281503 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-config-data\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.281550 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-kolla-config\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.384047 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgcbk\" (UniqueName: \"kubernetes.io/projected/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-kube-api-access-mgcbk\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.384143 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.384182 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.384216 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-config-data\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.384247 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-kolla-config\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.385222 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-kolla-config\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.385352 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-config-data\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.389206 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.389777 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.400566 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgcbk\" (UniqueName: \"kubernetes.io/projected/cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7-kube-api-access-mgcbk\") pod \"memcached-0\" (UID: \"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7\") " pod="openstack/memcached-0" Dec 04 06:28:11 crc kubenswrapper[4685]: I1204 06:28:11.575159 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 06:28:12 crc kubenswrapper[4685]: W1204 06:28:12.360235 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0cc031f_3f2c_4c6e_b46a_7513f15fd492.slice/crio-17fc92a4d1d69a4a00df048e3b0969f6b63ad7eb5394e0d1d5814d9ea96cbe99 WatchSource:0}: Error finding container 17fc92a4d1d69a4a00df048e3b0969f6b63ad7eb5394e0d1d5814d9ea96cbe99: Status 404 returned error can't find the container with id 17fc92a4d1d69a4a00df048e3b0969f6b63ad7eb5394e0d1d5814d9ea96cbe99 Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.139090 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p5754" event={"ID":"e0cc031f-3f2c-4c6e-b46a-7513f15fd492","Type":"ContainerStarted","Data":"17fc92a4d1d69a4a00df048e3b0969f6b63ad7eb5394e0d1d5814d9ea96cbe99"} Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.581856 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.583766 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.588256 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-w99ww" Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.598371 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.723735 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v254c\" (UniqueName: \"kubernetes.io/projected/9006f3bf-fc55-4204-8e07-df8cae636de4-kube-api-access-v254c\") pod \"kube-state-metrics-0\" (UID: \"9006f3bf-fc55-4204-8e07-df8cae636de4\") " pod="openstack/kube-state-metrics-0" Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.825354 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v254c\" (UniqueName: \"kubernetes.io/projected/9006f3bf-fc55-4204-8e07-df8cae636de4-kube-api-access-v254c\") pod \"kube-state-metrics-0\" (UID: \"9006f3bf-fc55-4204-8e07-df8cae636de4\") " pod="openstack/kube-state-metrics-0" Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.864677 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v254c\" (UniqueName: \"kubernetes.io/projected/9006f3bf-fc55-4204-8e07-df8cae636de4-kube-api-access-v254c\") pod \"kube-state-metrics-0\" (UID: \"9006f3bf-fc55-4204-8e07-df8cae636de4\") " pod="openstack/kube-state-metrics-0" Dec 04 06:28:13 crc kubenswrapper[4685]: I1204 06:28:13.932209 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 06:28:16 crc kubenswrapper[4685]: I1204 06:28:16.823949 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.371753 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z562k"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.376706 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.379621 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.380003 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-g86qc" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.380069 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.383877 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z562k"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.455079 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-wxmhr"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.457319 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.468424 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-wxmhr"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510217 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-run\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510263 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0269f-8ebe-48f0-81de-081da72da040-ovn-controller-tls-certs\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510289 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0269f-8ebe-48f0-81de-081da72da040-combined-ca-bundle\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510308 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-etc-ovs\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510328 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-scripts\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510358 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxpsx\" (UniqueName: \"kubernetes.io/projected/2db0269f-8ebe-48f0-81de-081da72da040-kube-api-access-gxpsx\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510396 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-log-ovn\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510429 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-lib\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510490 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-log\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510533 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-run-ovn\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510562 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q784\" (UniqueName: \"kubernetes.io/projected/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-kube-api-access-2q784\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db0269f-8ebe-48f0-81de-081da72da040-scripts\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.510615 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-run\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612247 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q784\" (UniqueName: \"kubernetes.io/projected/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-kube-api-access-2q784\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612316 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db0269f-8ebe-48f0-81de-081da72da040-scripts\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612385 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-run\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612440 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-run\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612466 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0269f-8ebe-48f0-81de-081da72da040-ovn-controller-tls-certs\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612509 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0269f-8ebe-48f0-81de-081da72da040-combined-ca-bundle\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612533 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-etc-ovs\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612557 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-scripts\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612611 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxpsx\" (UniqueName: \"kubernetes.io/projected/2db0269f-8ebe-48f0-81de-081da72da040-kube-api-access-gxpsx\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612648 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-log-ovn\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612687 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-lib\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612713 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-log\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.612773 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-run-ovn\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.614703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-run\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.614798 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-run\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.615778 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-run-ovn\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.617788 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-etc-ovs\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.618699 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db0269f-8ebe-48f0-81de-081da72da040-scripts\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.619160 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-lib\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.619210 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2db0269f-8ebe-48f0-81de-081da72da040-var-log-ovn\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.619310 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-var-log\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.620979 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-scripts\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.622087 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0269f-8ebe-48f0-81de-081da72da040-ovn-controller-tls-certs\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.632289 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0269f-8ebe-48f0-81de-081da72da040-combined-ca-bundle\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.647230 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxpsx\" (UniqueName: \"kubernetes.io/projected/2db0269f-8ebe-48f0-81de-081da72da040-kube-api-access-gxpsx\") pod \"ovn-controller-z562k\" (UID: \"2db0269f-8ebe-48f0-81de-081da72da040\") " pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.647496 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q784\" (UniqueName: \"kubernetes.io/projected/3fee23d2-2ddf-42e2-b600-5d0da74b77bb-kube-api-access-2q784\") pod \"ovn-controller-ovs-wxmhr\" (UID: \"3fee23d2-2ddf-42e2-b600-5d0da74b77bb\") " pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.717371 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.777901 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.811618 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.813045 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.817591 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-nvhc8" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.818667 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.818781 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.818942 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.819053 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.847467 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918397 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918477 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918501 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpxdn\" (UniqueName: \"kubernetes.io/projected/c99c06e8-97e8-44d4-b532-a2ca943d826a-kube-api-access-fpxdn\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918530 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c99c06e8-97e8-44d4-b532-a2ca943d826a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918560 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c99c06e8-97e8-44d4-b532-a2ca943d826a-config\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918583 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c99c06e8-97e8-44d4-b532-a2ca943d826a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918610 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:17 crc kubenswrapper[4685]: I1204 06:28:17.918631 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.020662 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c99c06e8-97e8-44d4-b532-a2ca943d826a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.020794 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c99c06e8-97e8-44d4-b532-a2ca943d826a-config\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.020853 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c99c06e8-97e8-44d4-b532-a2ca943d826a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.020944 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.021007 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.021162 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.021224 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c99c06e8-97e8-44d4-b532-a2ca943d826a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.021238 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.021876 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpxdn\" (UniqueName: \"kubernetes.io/projected/c99c06e8-97e8-44d4-b532-a2ca943d826a-kube-api-access-fpxdn\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.021727 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.022660 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c99c06e8-97e8-44d4-b532-a2ca943d826a-config\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.022980 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c99c06e8-97e8-44d4-b532-a2ca943d826a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.029266 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.029914 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.032933 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99c06e8-97e8-44d4-b532-a2ca943d826a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.044938 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpxdn\" (UniqueName: \"kubernetes.io/projected/c99c06e8-97e8-44d4-b532-a2ca943d826a-kube-api-access-fpxdn\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.053826 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c99c06e8-97e8-44d4-b532-a2ca943d826a\") " pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:18 crc kubenswrapper[4685]: I1204 06:28:18.149588 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.592611 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.595137 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.597621 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.597775 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-4mqft" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.597666 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.598348 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.599721 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677393 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l7qr\" (UniqueName: \"kubernetes.io/projected/78fd03f6-d85a-4348-a910-2cd71864b1fc-kube-api-access-9l7qr\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677501 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677541 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677615 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78fd03f6-d85a-4348-a910-2cd71864b1fc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677654 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677859 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78fd03f6-d85a-4348-a910-2cd71864b1fc-config\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.677940 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/78fd03f6-d85a-4348-a910-2cd71864b1fc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.678027 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781230 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781326 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78fd03f6-d85a-4348-a910-2cd71864b1fc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781447 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781486 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78fd03f6-d85a-4348-a910-2cd71864b1fc-config\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781513 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/78fd03f6-d85a-4348-a910-2cd71864b1fc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781551 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.781618 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l7qr\" (UniqueName: \"kubernetes.io/projected/78fd03f6-d85a-4348-a910-2cd71864b1fc-kube-api-access-9l7qr\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.782451 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/78fd03f6-d85a-4348-a910-2cd71864b1fc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.782809 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.783267 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78fd03f6-d85a-4348-a910-2cd71864b1fc-config\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.783679 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78fd03f6-d85a-4348-a910-2cd71864b1fc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.786749 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.787099 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.790175 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fd03f6-d85a-4348-a910-2cd71864b1fc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.808972 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.809423 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l7qr\" (UniqueName: \"kubernetes.io/projected/78fd03f6-d85a-4348-a910-2cd71864b1fc-kube-api-access-9l7qr\") pod \"ovsdbserver-nb-0\" (UID: \"78fd03f6-d85a-4348-a910-2cd71864b1fc\") " pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:20 crc kubenswrapper[4685]: I1204 06:28:20.928573 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:21 crc kubenswrapper[4685]: W1204 06:28:21.398170 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod777e7995_c510_4d30_bcf1_d479e7cfa432.slice/crio-ceecd5d116fa402caf1aeb215c6ca24d49f210c1fe001fadb07df03c65aec1de WatchSource:0}: Error finding container ceecd5d116fa402caf1aeb215c6ca24d49f210c1fe001fadb07df03c65aec1de: Status 404 returned error can't find the container with id ceecd5d116fa402caf1aeb215c6ca24d49f210c1fe001fadb07df03c65aec1de Dec 04 06:28:21 crc kubenswrapper[4685]: I1204 06:28:21.867264 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6wc8c"] Dec 04 06:28:22 crc kubenswrapper[4685]: I1204 06:28:22.227968 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"777e7995-c510-4d30-bcf1-d479e7cfa432","Type":"ContainerStarted","Data":"ceecd5d116fa402caf1aeb215c6ca24d49f210c1fe001fadb07df03c65aec1de"} Dec 04 06:28:22 crc kubenswrapper[4685]: W1204 06:28:22.309074 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f031df1_b162_4ace_8c47_021a7d1c9e61.slice/crio-0f3c328a3a803b3b509841b3a7dfc4b1847d2a6da6c15fa8ede55d931ac465c4 WatchSource:0}: Error finding container 0f3c328a3a803b3b509841b3a7dfc4b1847d2a6da6c15fa8ede55d931ac465c4: Status 404 returned error can't find the container with id 0f3c328a3a803b3b509841b3a7dfc4b1847d2a6da6c15fa8ede55d931ac465c4 Dec 04 06:28:22 crc kubenswrapper[4685]: E1204 06:28:22.366574 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 06:28:22 crc kubenswrapper[4685]: E1204 06:28:22.366946 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xvwkx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-ppmsn_openstack(db42e3fb-f38f-4892-bf85-de67095bb931): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:28:22 crc kubenswrapper[4685]: E1204 06:28:22.368647 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" podUID="db42e3fb-f38f-4892-bf85-de67095bb931" Dec 04 06:28:22 crc kubenswrapper[4685]: E1204 06:28:22.375478 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 06:28:22 crc kubenswrapper[4685]: E1204 06:28:22.375650 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-szvmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-lhd8w_openstack(be665a16-e393-4fe7-9f6b-c6412d555eb1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:28:22 crc kubenswrapper[4685]: E1204 06:28:22.379852 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" podUID="be665a16-e393-4fe7-9f6b-c6412d555eb1" Dec 04 06:28:22 crc kubenswrapper[4685]: I1204 06:28:22.854756 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 06:28:22 crc kubenswrapper[4685]: I1204 06:28:22.861813 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:28:22 crc kubenswrapper[4685]: I1204 06:28:22.994141 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.013648 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 06:28:23 crc kubenswrapper[4685]: W1204 06:28:23.026427 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod577f79bb_8577_412a_92e1_bf56244becce.slice/crio-a5832f9cb05d554d4c058c016e793b5992346088cb5504332015b35de1f0d974 WatchSource:0}: Error finding container a5832f9cb05d554d4c058c016e793b5992346088cb5504332015b35de1f0d974: Status 404 returned error can't find the container with id a5832f9cb05d554d4c058c016e793b5992346088cb5504332015b35de1f0d974 Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.141259 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:28:23 crc kubenswrapper[4685]: W1204 06:28:23.144179 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9006f3bf_fc55_4204_8e07_df8cae636de4.slice/crio-d429a5993dc4cc176b1e7f81a6885eab50663f2eb8b24b9c593038f0a0029313 WatchSource:0}: Error finding container d429a5993dc4cc176b1e7f81a6885eab50663f2eb8b24b9c593038f0a0029313: Status 404 returned error can't find the container with id d429a5993dc4cc176b1e7f81a6885eab50663f2eb8b24b9c593038f0a0029313 Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.256997 4685 generic.go:334] "Generic (PLEG): container finished" podID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerID="be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d" exitCode=0 Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.257152 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p5754" event={"ID":"e0cc031f-3f2c-4c6e-b46a-7513f15fd492","Type":"ContainerDied","Data":"be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.260101 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.267450 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z562k"] Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.279876 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ff4ac6f-e715-439b-8b4a-2b2715198dab","Type":"ContainerStarted","Data":"d3162bdd921ea7ebf8c3544ad94fcdb99d442eed9134d0d62ca53d7fc18493fd"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.289397 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7","Type":"ContainerStarted","Data":"1449a0e3105931d66011a0fdfe528c50b3ef8be7e12f5e9da5b50d3e8e02c0e9"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.306684 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"577f79bb-8577-412a-92e1-bf56244becce","Type":"ContainerStarted","Data":"a5832f9cb05d554d4c058c016e793b5992346088cb5504332015b35de1f0d974"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.308492 4685 generic.go:334] "Generic (PLEG): container finished" podID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerID="3f7646ff4e72435b72c62993daf80bcfbd0f281b3d771c84f1c67022a2eaa3a9" exitCode=0 Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.308546 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" event={"ID":"6f031df1-b162-4ace-8c47-021a7d1c9e61","Type":"ContainerDied","Data":"3f7646ff4e72435b72c62993daf80bcfbd0f281b3d771c84f1c67022a2eaa3a9"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.308566 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" event={"ID":"6f031df1-b162-4ace-8c47-021a7d1c9e61","Type":"ContainerStarted","Data":"0f3c328a3a803b3b509841b3a7dfc4b1847d2a6da6c15fa8ede55d931ac465c4"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.325347 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9006f3bf-fc55-4204-8e07-df8cae636de4","Type":"ContainerStarted","Data":"d429a5993dc4cc176b1e7f81a6885eab50663f2eb8b24b9c593038f0a0029313"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.343301 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"536ebc06-2d82-419d-ac24-957fbc6734c9","Type":"ContainerStarted","Data":"6f65eb895482220b0479cc91e2c7501e8850605a70a6c5c6be134a80642ff977"} Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.427943 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.780195 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-wxmhr"] Dec 04 06:28:23 crc kubenswrapper[4685]: W1204 06:28:23.796070 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fee23d2_2ddf_42e2_b600_5d0da74b77bb.slice/crio-e11fd8482994dd41b9936c19ac589c3e06282dcc4203e504ca346b48346532e0 WatchSource:0}: Error finding container e11fd8482994dd41b9936c19ac589c3e06282dcc4203e504ca346b48346532e0: Status 404 returned error can't find the container with id e11fd8482994dd41b9936c19ac589c3e06282dcc4203e504ca346b48346532e0 Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.842026 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.938745 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.966256 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szvmn\" (UniqueName: \"kubernetes.io/projected/be665a16-e393-4fe7-9f6b-c6412d555eb1-kube-api-access-szvmn\") pod \"be665a16-e393-4fe7-9f6b-c6412d555eb1\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.966343 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be665a16-e393-4fe7-9f6b-c6412d555eb1-config\") pod \"be665a16-e393-4fe7-9f6b-c6412d555eb1\" (UID: \"be665a16-e393-4fe7-9f6b-c6412d555eb1\") " Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.967426 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be665a16-e393-4fe7-9f6b-c6412d555eb1-config" (OuterVolumeSpecName: "config") pod "be665a16-e393-4fe7-9f6b-c6412d555eb1" (UID: "be665a16-e393-4fe7-9f6b-c6412d555eb1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:23 crc kubenswrapper[4685]: I1204 06:28:23.980287 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be665a16-e393-4fe7-9f6b-c6412d555eb1-kube-api-access-szvmn" (OuterVolumeSpecName: "kube-api-access-szvmn") pod "be665a16-e393-4fe7-9f6b-c6412d555eb1" (UID: "be665a16-e393-4fe7-9f6b-c6412d555eb1"). InnerVolumeSpecName "kube-api-access-szvmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.067670 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-config\") pod \"db42e3fb-f38f-4892-bf85-de67095bb931\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.067937 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-dns-svc\") pod \"db42e3fb-f38f-4892-bf85-de67095bb931\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.068007 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvwkx\" (UniqueName: \"kubernetes.io/projected/db42e3fb-f38f-4892-bf85-de67095bb931-kube-api-access-xvwkx\") pod \"db42e3fb-f38f-4892-bf85-de67095bb931\" (UID: \"db42e3fb-f38f-4892-bf85-de67095bb931\") " Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.068338 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-config" (OuterVolumeSpecName: "config") pod "db42e3fb-f38f-4892-bf85-de67095bb931" (UID: "db42e3fb-f38f-4892-bf85-de67095bb931"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.068700 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db42e3fb-f38f-4892-bf85-de67095bb931" (UID: "db42e3fb-f38f-4892-bf85-de67095bb931"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.068941 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.068964 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szvmn\" (UniqueName: \"kubernetes.io/projected/be665a16-e393-4fe7-9f6b-c6412d555eb1-kube-api-access-szvmn\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.068978 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be665a16-e393-4fe7-9f6b-c6412d555eb1-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.073996 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db42e3fb-f38f-4892-bf85-de67095bb931-kube-api-access-xvwkx" (OuterVolumeSpecName: "kube-api-access-xvwkx") pod "db42e3fb-f38f-4892-bf85-de67095bb931" (UID: "db42e3fb-f38f-4892-bf85-de67095bb931"). InnerVolumeSpecName "kube-api-access-xvwkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.172759 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db42e3fb-f38f-4892-bf85-de67095bb931-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.172794 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvwkx\" (UniqueName: \"kubernetes.io/projected/db42e3fb-f38f-4892-bf85-de67095bb931-kube-api-access-xvwkx\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.359839 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wxmhr" event={"ID":"3fee23d2-2ddf-42e2-b600-5d0da74b77bb","Type":"ContainerStarted","Data":"e11fd8482994dd41b9936c19ac589c3e06282dcc4203e504ca346b48346532e0"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.370490 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" event={"ID":"be665a16-e393-4fe7-9f6b-c6412d555eb1","Type":"ContainerDied","Data":"10d0e5af7121121d0f8533e1e7dca45c2a9ebc540f9073b3f6a6fedaa6375e24"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.370534 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lhd8w" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.376714 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"78fd03f6-d85a-4348-a910-2cd71864b1fc","Type":"ContainerStarted","Data":"62ed0e4bea1c4411c801dc6c1d3e6f948b5621e1e5d93f7162a77ed556f5e549"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.380615 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c99c06e8-97e8-44d4-b532-a2ca943d826a","Type":"ContainerStarted","Data":"edfff799d228bc0d3c374e01d450bebd32c1a4c7d186ad636b9f973f292f9340"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.384204 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p5754" event={"ID":"e0cc031f-3f2c-4c6e-b46a-7513f15fd492","Type":"ContainerStarted","Data":"a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.384314 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.387868 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k" event={"ID":"2db0269f-8ebe-48f0-81de-081da72da040","Type":"ContainerStarted","Data":"703ffa29655043e01d02072517308b9a456a3a1252204d94aff89bc0c6639c74"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.391505 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" event={"ID":"db42e3fb-f38f-4892-bf85-de67095bb931","Type":"ContainerDied","Data":"8df0a712864f5e78b675cf41ae2dff4482e9bf416d0544b2b81f0a7571c109e5"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.391797 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ppmsn" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.401209 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" event={"ID":"6f031df1-b162-4ace-8c47-021a7d1c9e61","Type":"ContainerStarted","Data":"1fd45d68951d01729b841414b884c7a72b5937aa612b095d00f9b4c1a4711998"} Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.403317 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.424924 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-p5754" podStartSLOduration=8.351232737 podStartE2EDuration="18.424887919s" podCreationTimestamp="2025-12-04 06:28:06 +0000 UTC" firstStartedPulling="2025-12-04 06:28:12.372077677 +0000 UTC m=+1029.560308462" lastFinishedPulling="2025-12-04 06:28:22.445732869 +0000 UTC m=+1039.633963644" observedRunningTime="2025-12-04 06:28:24.402875565 +0000 UTC m=+1041.591106360" watchObservedRunningTime="2025-12-04 06:28:24.424887919 +0000 UTC m=+1041.613118714" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.524713 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lhd8w"] Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.540598 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lhd8w"] Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.542453 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" podStartSLOduration=18.085438681 podStartE2EDuration="18.542424586s" podCreationTimestamp="2025-12-04 06:28:06 +0000 UTC" firstStartedPulling="2025-12-04 06:28:22.323189824 +0000 UTC m=+1039.511420599" lastFinishedPulling="2025-12-04 06:28:22.780175729 +0000 UTC m=+1039.968406504" observedRunningTime="2025-12-04 06:28:24.473920899 +0000 UTC m=+1041.662151674" watchObservedRunningTime="2025-12-04 06:28:24.542424586 +0000 UTC m=+1041.730655361" Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.567204 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ppmsn"] Dec 04 06:28:24 crc kubenswrapper[4685]: I1204 06:28:24.573207 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ppmsn"] Dec 04 06:28:25 crc kubenswrapper[4685]: I1204 06:28:25.138031 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be665a16-e393-4fe7-9f6b-c6412d555eb1" path="/var/lib/kubelet/pods/be665a16-e393-4fe7-9f6b-c6412d555eb1/volumes" Dec 04 06:28:25 crc kubenswrapper[4685]: I1204 06:28:25.139316 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db42e3fb-f38f-4892-bf85-de67095bb931" path="/var/lib/kubelet/pods/db42e3fb-f38f-4892-bf85-de67095bb931/volumes" Dec 04 06:28:31 crc kubenswrapper[4685]: I1204 06:28:31.703596 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:32 crc kubenswrapper[4685]: I1204 06:28:32.019815 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:32 crc kubenswrapper[4685]: I1204 06:28:32.086189 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p5754"] Dec 04 06:28:32 crc kubenswrapper[4685]: I1204 06:28:32.468260 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-p5754" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerName="dnsmasq-dns" containerID="cri-o://a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7" gracePeriod=10 Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.084742 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.154746 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlzt9\" (UniqueName: \"kubernetes.io/projected/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-kube-api-access-zlzt9\") pod \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.154823 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-dns-svc\") pod \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.154898 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-config\") pod \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\" (UID: \"e0cc031f-3f2c-4c6e-b46a-7513f15fd492\") " Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.202731 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-kube-api-access-zlzt9" (OuterVolumeSpecName: "kube-api-access-zlzt9") pod "e0cc031f-3f2c-4c6e-b46a-7513f15fd492" (UID: "e0cc031f-3f2c-4c6e-b46a-7513f15fd492"). InnerVolumeSpecName "kube-api-access-zlzt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.259168 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlzt9\" (UniqueName: \"kubernetes.io/projected/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-kube-api-access-zlzt9\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.483324 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9006f3bf-fc55-4204-8e07-df8cae636de4","Type":"ContainerStarted","Data":"c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a"} Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.484995 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.489305 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"78fd03f6-d85a-4348-a910-2cd71864b1fc","Type":"ContainerStarted","Data":"c99a54291f08b3dfcfefe78ba61400f264be9cb46c8f41db8b1452521dd2278a"} Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.499034 4685 generic.go:334] "Generic (PLEG): container finished" podID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerID="a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7" exitCode=0 Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.499125 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p5754" event={"ID":"e0cc031f-3f2c-4c6e-b46a-7513f15fd492","Type":"ContainerDied","Data":"a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7"} Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.499170 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p5754" event={"ID":"e0cc031f-3f2c-4c6e-b46a-7513f15fd492","Type":"ContainerDied","Data":"17fc92a4d1d69a4a00df048e3b0969f6b63ad7eb5394e0d1d5814d9ea96cbe99"} Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.499195 4685 scope.go:117] "RemoveContainer" containerID="a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.499392 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p5754" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.506022 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7","Type":"ContainerStarted","Data":"2d9c8bd86c1429bf4f610e561a6798ee4d0af8bf7df01c6ebab7b28ba0f5c3d4"} Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.506530 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.514047 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.105788706 podStartE2EDuration="20.514020982s" podCreationTimestamp="2025-12-04 06:28:13 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.151779484 +0000 UTC m=+1040.340010249" lastFinishedPulling="2025-12-04 06:28:32.56001175 +0000 UTC m=+1049.748242525" observedRunningTime="2025-12-04 06:28:33.511828176 +0000 UTC m=+1050.700058951" watchObservedRunningTime="2025-12-04 06:28:33.514020982 +0000 UTC m=+1050.702251757" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.515435 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"577f79bb-8577-412a-92e1-bf56244becce","Type":"ContainerStarted","Data":"25e2de64f1e9e9fb40a6e9f201b5cfbb88313310889a322420f398eabd38452f"} Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.527200 4685 scope.go:117] "RemoveContainer" containerID="be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.572000 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.115438114 podStartE2EDuration="22.571971426s" podCreationTimestamp="2025-12-04 06:28:11 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.026931311 +0000 UTC m=+1040.215162086" lastFinishedPulling="2025-12-04 06:28:31.483464623 +0000 UTC m=+1048.671695398" observedRunningTime="2025-12-04 06:28:33.561549336 +0000 UTC m=+1050.749780121" watchObservedRunningTime="2025-12-04 06:28:33.571971426 +0000 UTC m=+1050.760202211" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.587274 4685 scope.go:117] "RemoveContainer" containerID="a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7" Dec 04 06:28:33 crc kubenswrapper[4685]: E1204 06:28:33.588448 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7\": container with ID starting with a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7 not found: ID does not exist" containerID="a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.588525 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7"} err="failed to get container status \"a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7\": rpc error: code = NotFound desc = could not find container \"a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7\": container with ID starting with a8fed2102c2045ef8a8a2f3c39b23e7a8e871b796382f6e868d54baa0cbef3a7 not found: ID does not exist" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.592376 4685 scope.go:117] "RemoveContainer" containerID="be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d" Dec 04 06:28:33 crc kubenswrapper[4685]: E1204 06:28:33.594354 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d\": container with ID starting with be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d not found: ID does not exist" containerID="be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.594434 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d"} err="failed to get container status \"be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d\": rpc error: code = NotFound desc = could not find container \"be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d\": container with ID starting with be6e3d25f2df73dfedcd8928017d5de68e9de5253bb2dabf48c9bdf9e9fc347d not found: ID does not exist" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.685514 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-config" (OuterVolumeSpecName: "config") pod "e0cc031f-3f2c-4c6e-b46a-7513f15fd492" (UID: "e0cc031f-3f2c-4c6e-b46a-7513f15fd492"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.771321 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.907588 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e0cc031f-3f2c-4c6e-b46a-7513f15fd492" (UID: "e0cc031f-3f2c-4c6e-b46a-7513f15fd492"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:33 crc kubenswrapper[4685]: I1204 06:28:33.974851 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0cc031f-3f2c-4c6e-b46a-7513f15fd492-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.164287 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p5754"] Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.176000 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p5754"] Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.536610 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"536ebc06-2d82-419d-ac24-957fbc6734c9","Type":"ContainerStarted","Data":"d2820b911c8b8ecdd31600f26f5440dea701231a129c2ec6d7a0003956a569f3"} Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.556154 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c99c06e8-97e8-44d4-b532-a2ca943d826a","Type":"ContainerStarted","Data":"4e7b6a9c416f9bac2d84fca0d22b55ae2c48485489efd4eae13defe6ccfbf054"} Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.564004 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ff4ac6f-e715-439b-8b4a-2b2715198dab","Type":"ContainerStarted","Data":"a1c9f7e2cb752c2aae9f83d4ccd671a7be6e02092fac5863bf3f397693e89979"} Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.578642 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k" event={"ID":"2db0269f-8ebe-48f0-81de-081da72da040","Type":"ContainerStarted","Data":"18ae335aaa63ece67541ae6e6f3b2515214ada612e2ffeeda0a674f6c1ccaf54"} Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.579902 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-z562k" Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.587113 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"777e7995-c510-4d30-bcf1-d479e7cfa432","Type":"ContainerStarted","Data":"d20cedfa26345e0d3ffb056841c394db2b49e8dab26dc2b4f19d7034f3ab834f"} Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.594353 4685 generic.go:334] "Generic (PLEG): container finished" podID="3fee23d2-2ddf-42e2-b600-5d0da74b77bb" containerID="2ee323976282a6206def11c5fe76b4122ed48f05082dfdf75a96d8843cb1e143" exitCode=0 Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.594728 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wxmhr" event={"ID":"3fee23d2-2ddf-42e2-b600-5d0da74b77bb","Type":"ContainerDied","Data":"2ee323976282a6206def11c5fe76b4122ed48f05082dfdf75a96d8843cb1e143"} Dec 04 06:28:34 crc kubenswrapper[4685]: I1204 06:28:34.644561 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z562k" podStartSLOduration=9.066626611 podStartE2EDuration="17.644542285s" podCreationTimestamp="2025-12-04 06:28:17 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.277801644 +0000 UTC m=+1040.466032419" lastFinishedPulling="2025-12-04 06:28:31.855717318 +0000 UTC m=+1049.043948093" observedRunningTime="2025-12-04 06:28:34.639421833 +0000 UTC m=+1051.827652598" watchObservedRunningTime="2025-12-04 06:28:34.644542285 +0000 UTC m=+1051.832773060" Dec 04 06:28:35 crc kubenswrapper[4685]: I1204 06:28:35.137231 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" path="/var/lib/kubelet/pods/e0cc031f-3f2c-4c6e-b46a-7513f15fd492/volumes" Dec 04 06:28:35 crc kubenswrapper[4685]: I1204 06:28:35.612677 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wxmhr" event={"ID":"3fee23d2-2ddf-42e2-b600-5d0da74b77bb","Type":"ContainerStarted","Data":"6c21efbec657e1d2c84ba314f0247a7b6149246c943a3c355cd6e3e1f098891e"} Dec 04 06:28:35 crc kubenswrapper[4685]: I1204 06:28:35.613069 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wxmhr" event={"ID":"3fee23d2-2ddf-42e2-b600-5d0da74b77bb","Type":"ContainerStarted","Data":"80fa5ec4cdf8383d24213b4fce7700ab4ed85103350a62de3caf969f9c854153"} Dec 04 06:28:35 crc kubenswrapper[4685]: I1204 06:28:35.641516 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-wxmhr" podStartSLOduration=10.801289788 podStartE2EDuration="18.641484925s" podCreationTimestamp="2025-12-04 06:28:17 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.807329747 +0000 UTC m=+1040.995560522" lastFinishedPulling="2025-12-04 06:28:31.647524884 +0000 UTC m=+1048.835755659" observedRunningTime="2025-12-04 06:28:35.634870748 +0000 UTC m=+1052.823101523" watchObservedRunningTime="2025-12-04 06:28:35.641484925 +0000 UTC m=+1052.829715700" Dec 04 06:28:36 crc kubenswrapper[4685]: I1204 06:28:36.623050 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:36 crc kubenswrapper[4685]: I1204 06:28:36.623653 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:28:37 crc kubenswrapper[4685]: I1204 06:28:37.634555 4685 generic.go:334] "Generic (PLEG): container finished" podID="577f79bb-8577-412a-92e1-bf56244becce" containerID="25e2de64f1e9e9fb40a6e9f201b5cfbb88313310889a322420f398eabd38452f" exitCode=0 Dec 04 06:28:37 crc kubenswrapper[4685]: I1204 06:28:37.634627 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"577f79bb-8577-412a-92e1-bf56244becce","Type":"ContainerDied","Data":"25e2de64f1e9e9fb40a6e9f201b5cfbb88313310889a322420f398eabd38452f"} Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.650221 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"577f79bb-8577-412a-92e1-bf56244becce","Type":"ContainerStarted","Data":"9b94d3db562f81783b365951dfc219134a1e4658217646c8a8a4becec8ce61f9"} Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.651902 4685 generic.go:334] "Generic (PLEG): container finished" podID="536ebc06-2d82-419d-ac24-957fbc6734c9" containerID="d2820b911c8b8ecdd31600f26f5440dea701231a129c2ec6d7a0003956a569f3" exitCode=0 Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.652031 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"536ebc06-2d82-419d-ac24-957fbc6734c9","Type":"ContainerDied","Data":"d2820b911c8b8ecdd31600f26f5440dea701231a129c2ec6d7a0003956a569f3"} Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.654775 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"78fd03f6-d85a-4348-a910-2cd71864b1fc","Type":"ContainerStarted","Data":"e58eb9b57ea93629c89f7ff947e85b0f3f825a9408d78bc32c44924d6efcde48"} Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.658561 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c99c06e8-97e8-44d4-b532-a2ca943d826a","Type":"ContainerStarted","Data":"a14fffba9687104781fc5c0077ae161b90cdee6ac5873cdebf7fd8b00584597e"} Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.691562 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.349019342 podStartE2EDuration="30.691533413s" podCreationTimestamp="2025-12-04 06:28:08 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.031066463 +0000 UTC m=+1040.219297238" lastFinishedPulling="2025-12-04 06:28:32.373580524 +0000 UTC m=+1049.561811309" observedRunningTime="2025-12-04 06:28:38.687945146 +0000 UTC m=+1055.876175931" watchObservedRunningTime="2025-12-04 06:28:38.691533413 +0000 UTC m=+1055.879764228" Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.735256 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.739056897 podStartE2EDuration="19.735225694s" podCreationTimestamp="2025-12-04 06:28:19 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.497029356 +0000 UTC m=+1040.685260131" lastFinishedPulling="2025-12-04 06:28:37.493198153 +0000 UTC m=+1054.681428928" observedRunningTime="2025-12-04 06:28:38.718927898 +0000 UTC m=+1055.907158723" watchObservedRunningTime="2025-12-04 06:28:38.735225694 +0000 UTC m=+1055.923456499" Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.774311 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.574920784 podStartE2EDuration="22.774241285s" podCreationTimestamp="2025-12-04 06:28:16 +0000 UTC" firstStartedPulling="2025-12-04 06:28:23.27835185 +0000 UTC m=+1040.466582635" lastFinishedPulling="2025-12-04 06:28:37.477672361 +0000 UTC m=+1054.665903136" observedRunningTime="2025-12-04 06:28:38.754791326 +0000 UTC m=+1055.943022171" watchObservedRunningTime="2025-12-04 06:28:38.774241285 +0000 UTC m=+1055.962472100" Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.928754 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:38 crc kubenswrapper[4685]: I1204 06:28:38.980074 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.150125 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.195763 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.680500 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"536ebc06-2d82-419d-ac24-957fbc6734c9","Type":"ContainerStarted","Data":"bd72df6f72b1d84fe81f57ccb422d5cb359429f279e567cd46e210ec5567aa38"} Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.680573 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.683164 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.717211 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.932217484 podStartE2EDuration="30.717186127s" podCreationTimestamp="2025-12-04 06:28:09 +0000 UTC" firstStartedPulling="2025-12-04 06:28:22.862416886 +0000 UTC m=+1040.050647661" lastFinishedPulling="2025-12-04 06:28:31.647385529 +0000 UTC m=+1048.835616304" observedRunningTime="2025-12-04 06:28:39.711938721 +0000 UTC m=+1056.900169536" watchObservedRunningTime="2025-12-04 06:28:39.717186127 +0000 UTC m=+1056.905416912" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.744548 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 06:28:39 crc kubenswrapper[4685]: I1204 06:28:39.755666 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.010625 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.011092 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.045309 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gjg6f"] Dec 04 06:28:40 crc kubenswrapper[4685]: E1204 06:28:40.045702 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerName="dnsmasq-dns" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.045719 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerName="dnsmasq-dns" Dec 04 06:28:40 crc kubenswrapper[4685]: E1204 06:28:40.045743 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerName="init" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.045751 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerName="init" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.045937 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cc031f-3f2c-4c6e-b46a-7513f15fd492" containerName="dnsmasq-dns" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.050058 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.053859 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.067166 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gjg6f"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.112187 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7njcr"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.113292 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.120047 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.122878 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7njcr"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd2s7\" (UniqueName: \"kubernetes.io/projected/29b767dd-48fe-49d1-969a-e0347cf9a2c4-kube-api-access-pd2s7\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196685 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhrw9\" (UniqueName: \"kubernetes.io/projected/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-kube-api-access-qhrw9\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-config\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196789 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/29b767dd-48fe-49d1-969a-e0347cf9a2c4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196815 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/29b767dd-48fe-49d1-969a-e0347cf9a2c4-ovn-rundir\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196876 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196894 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b767dd-48fe-49d1-969a-e0347cf9a2c4-combined-ca-bundle\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196912 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/29b767dd-48fe-49d1-969a-e0347cf9a2c4-ovs-rundir\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196946 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29b767dd-48fe-49d1-969a-e0347cf9a2c4-config\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.196991 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.198527 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gjg6f"] Dec 04 06:28:40 crc kubenswrapper[4685]: E1204 06:28:40.199243 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-qhrw9 ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" podUID="7787b5e9-b8c6-45bd-85c3-85aa240eb3b3" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.234134 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-czlv2"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.236653 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.241683 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.245468 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.246847 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.249740 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.249927 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.249950 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-95297" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.258394 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-czlv2"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.260822 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.267707 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299097 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299148 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299168 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299195 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd2s7\" (UniqueName: \"kubernetes.io/projected/29b767dd-48fe-49d1-969a-e0347cf9a2c4-kube-api-access-pd2s7\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299224 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299242 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-dns-svc\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299261 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhrw9\" (UniqueName: \"kubernetes.io/projected/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-kube-api-access-qhrw9\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299287 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-scripts\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299305 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-config\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299324 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grtzt\" (UniqueName: \"kubernetes.io/projected/313ed371-bc49-447b-b609-7ec6a346e3f6-kube-api-access-grtzt\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299357 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-config\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299376 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/29b767dd-48fe-49d1-969a-e0347cf9a2c4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299396 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/29b767dd-48fe-49d1-969a-e0347cf9a2c4-ovn-rundir\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299433 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-config\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299462 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299478 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b767dd-48fe-49d1-969a-e0347cf9a2c4-combined-ca-bundle\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/29b767dd-48fe-49d1-969a-e0347cf9a2c4-ovs-rundir\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299524 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29b767dd-48fe-49d1-969a-e0347cf9a2c4-config\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299541 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299559 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299591 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.299614 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgfx8\" (UniqueName: \"kubernetes.io/projected/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-kube-api-access-fgfx8\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.300647 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.301799 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-config\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.302376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.304318 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/29b767dd-48fe-49d1-969a-e0347cf9a2c4-ovs-rundir\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.304601 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/29b767dd-48fe-49d1-969a-e0347cf9a2c4-ovn-rundir\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.305042 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29b767dd-48fe-49d1-969a-e0347cf9a2c4-config\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.310824 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/29b767dd-48fe-49d1-969a-e0347cf9a2c4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.319147 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b767dd-48fe-49d1-969a-e0347cf9a2c4-combined-ca-bundle\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.321922 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd2s7\" (UniqueName: \"kubernetes.io/projected/29b767dd-48fe-49d1-969a-e0347cf9a2c4-kube-api-access-pd2s7\") pod \"ovn-controller-metrics-7njcr\" (UID: \"29b767dd-48fe-49d1-969a-e0347cf9a2c4\") " pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.334255 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhrw9\" (UniqueName: \"kubernetes.io/projected/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-kube-api-access-qhrw9\") pod \"dnsmasq-dns-6bc7876d45-gjg6f\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.401804 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-config\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.401933 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.401967 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402006 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402037 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgfx8\" (UniqueName: \"kubernetes.io/projected/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-kube-api-access-fgfx8\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402074 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402098 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402139 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402158 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-dns-svc\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402192 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-scripts\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402214 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grtzt\" (UniqueName: \"kubernetes.io/projected/313ed371-bc49-447b-b609-7ec6a346e3f6-kube-api-access-grtzt\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402246 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-config\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.402943 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-config\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.403429 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-config\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.403562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.403798 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.404048 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-scripts\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.409185 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.409709 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.409759 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-dns-svc\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.410423 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.410974 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.428567 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgfx8\" (UniqueName: \"kubernetes.io/projected/53bf9890-b3a9-4102-9b2f-f7711e7dbe70-kube-api-access-fgfx8\") pod \"ovn-northd-0\" (UID: \"53bf9890-b3a9-4102-9b2f-f7711e7dbe70\") " pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.431136 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7njcr" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.434359 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grtzt\" (UniqueName: \"kubernetes.io/projected/313ed371-bc49-447b-b609-7ec6a346e3f6-kube-api-access-grtzt\") pod \"dnsmasq-dns-8554648995-czlv2\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.559889 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.595681 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.689298 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.710119 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.813245 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-ovsdbserver-sb\") pod \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.814082 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-config\") pod \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.813908 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3" (UID: "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.814800 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-config" (OuterVolumeSpecName: "config") pod "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3" (UID: "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.815701 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhrw9\" (UniqueName: \"kubernetes.io/projected/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-kube-api-access-qhrw9\") pod \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.815741 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-dns-svc\") pod \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\" (UID: \"7787b5e9-b8c6-45bd-85c3-85aa240eb3b3\") " Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.817134 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.817159 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.818198 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3" (UID: "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.823794 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-kube-api-access-qhrw9" (OuterVolumeSpecName: "kube-api-access-qhrw9") pod "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3" (UID: "7787b5e9-b8c6-45bd-85c3-85aa240eb3b3"). InnerVolumeSpecName "kube-api-access-qhrw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.920003 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhrw9\" (UniqueName: \"kubernetes.io/projected/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-kube-api-access-qhrw9\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.920055 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:40 crc kubenswrapper[4685]: I1204 06:28:40.962625 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7njcr"] Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.091670 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-czlv2"] Dec 04 06:28:41 crc kubenswrapper[4685]: W1204 06:28:41.100255 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod313ed371_bc49_447b_b609_7ec6a346e3f6.slice/crio-ff0cc635cf11533eeb116e2abeb2e7ea8c2856801c827b6f693e5c9fdfd9d066 WatchSource:0}: Error finding container ff0cc635cf11533eeb116e2abeb2e7ea8c2856801c827b6f693e5c9fdfd9d066: Status 404 returned error can't find the container with id ff0cc635cf11533eeb116e2abeb2e7ea8c2856801c827b6f693e5c9fdfd9d066 Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.160392 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 06:28:41 crc kubenswrapper[4685]: W1204 06:28:41.184666 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53bf9890_b3a9_4102_9b2f_f7711e7dbe70.slice/crio-737d89498bf6be4e8cba7e6d04b8ac6f66e84d77f8b94347b0259c51fe416a82 WatchSource:0}: Error finding container 737d89498bf6be4e8cba7e6d04b8ac6f66e84d77f8b94347b0259c51fe416a82: Status 404 returned error can't find the container with id 737d89498bf6be4e8cba7e6d04b8ac6f66e84d77f8b94347b0259c51fe416a82 Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.250491 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.250533 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.578090 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.697876 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"53bf9890-b3a9-4102-9b2f-f7711e7dbe70","Type":"ContainerStarted","Data":"737d89498bf6be4e8cba7e6d04b8ac6f66e84d77f8b94347b0259c51fe416a82"} Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.703897 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7njcr" event={"ID":"29b767dd-48fe-49d1-969a-e0347cf9a2c4","Type":"ContainerStarted","Data":"fff237324e362b0b4a812f86b7df59a9ce21f1c290262d7fa402c5ba9527409d"} Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.703951 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7njcr" event={"ID":"29b767dd-48fe-49d1-969a-e0347cf9a2c4","Type":"ContainerStarted","Data":"c086cda535d48c707669b1b59f5ebeea6f3a106df8f27739deac363d31655dad"} Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.705977 4685 generic.go:334] "Generic (PLEG): container finished" podID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerID="335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb" exitCode=0 Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.706130 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gjg6f" Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.706387 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-czlv2" event={"ID":"313ed371-bc49-447b-b609-7ec6a346e3f6","Type":"ContainerDied","Data":"335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb"} Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.706454 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-czlv2" event={"ID":"313ed371-bc49-447b-b609-7ec6a346e3f6","Type":"ContainerStarted","Data":"ff0cc635cf11533eeb116e2abeb2e7ea8c2856801c827b6f693e5c9fdfd9d066"} Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.726400 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7njcr" podStartSLOduration=1.726382461 podStartE2EDuration="1.726382461s" podCreationTimestamp="2025-12-04 06:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:28:41.724826845 +0000 UTC m=+1058.913057620" watchObservedRunningTime="2025-12-04 06:28:41.726382461 +0000 UTC m=+1058.914613236" Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.836014 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gjg6f"] Dec 04 06:28:41 crc kubenswrapper[4685]: I1204 06:28:41.862238 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gjg6f"] Dec 04 06:28:43 crc kubenswrapper[4685]: I1204 06:28:43.140595 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7787b5e9-b8c6-45bd-85c3-85aa240eb3b3" path="/var/lib/kubelet/pods/7787b5e9-b8c6-45bd-85c3-85aa240eb3b3/volumes" Dec 04 06:28:43 crc kubenswrapper[4685]: I1204 06:28:43.941641 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-czlv2"] Dec 04 06:28:43 crc kubenswrapper[4685]: I1204 06:28:43.949583 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 06:28:43 crc kubenswrapper[4685]: I1204 06:28:43.998823 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-6bsx8"] Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.000128 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.010182 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-6bsx8"] Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.097182 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.097239 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kfxh\" (UniqueName: \"kubernetes.io/projected/de4c5c56-929c-47b1-a488-f9eea7edeccc-kube-api-access-9kfxh\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.097277 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.097311 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-config\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.097427 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.199186 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.199291 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.199313 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kfxh\" (UniqueName: \"kubernetes.io/projected/de4c5c56-929c-47b1-a488-f9eea7edeccc-kube-api-access-9kfxh\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.199335 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.199357 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-config\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.200563 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.200899 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-config\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.201113 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.202228 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.218912 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kfxh\" (UniqueName: \"kubernetes.io/projected/de4c5c56-929c-47b1-a488-f9eea7edeccc-kube-api-access-9kfxh\") pod \"dnsmasq-dns-b8fbc5445-6bsx8\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.344327 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:44 crc kubenswrapper[4685]: I1204 06:28:44.804196 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-6bsx8"] Dec 04 06:28:44 crc kubenswrapper[4685]: W1204 06:28:44.810649 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde4c5c56_929c_47b1_a488_f9eea7edeccc.slice/crio-8993fe3de7909e5d68edd6aa502520fae46a0622c78d34b6af3405418aee427a WatchSource:0}: Error finding container 8993fe3de7909e5d68edd6aa502520fae46a0622c78d34b6af3405418aee427a: Status 404 returned error can't find the container with id 8993fe3de7909e5d68edd6aa502520fae46a0622c78d34b6af3405418aee427a Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.191792 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.200502 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.202897 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.205996 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.205996 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-2ltlx" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.206054 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.207281 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.324495 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af46303b-ea28-4ab3-acb7-56a16afa6afd-cache\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.324569 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af46303b-ea28-4ab3-acb7-56a16afa6afd-lock\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.324808 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.325112 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.325155 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r2rp\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-kube-api-access-7r2rp\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.426532 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.426638 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r2rp\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-kube-api-access-7r2rp\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.426751 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af46303b-ea28-4ab3-acb7-56a16afa6afd-cache\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.426818 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af46303b-ea28-4ab3-acb7-56a16afa6afd-lock\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: E1204 06:28:45.426903 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 06:28:45 crc kubenswrapper[4685]: E1204 06:28:45.426947 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.426960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: E1204 06:28:45.427050 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift podName:af46303b-ea28-4ab3-acb7-56a16afa6afd nodeName:}" failed. No retries permitted until 2025-12-04 06:28:45.927010085 +0000 UTC m=+1063.115240900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift") pod "swift-storage-0" (UID: "af46303b-ea28-4ab3-acb7-56a16afa6afd") : configmap "swift-ring-files" not found Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.427581 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.427593 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/af46303b-ea28-4ab3-acb7-56a16afa6afd-cache\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.427771 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/af46303b-ea28-4ab3-acb7-56a16afa6afd-lock\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.458019 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r2rp\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-kube-api-access-7r2rp\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.491740 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.495069 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9bq7c"] Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.496324 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.503471 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.510220 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.510458 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.512936 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9bq7c"] Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.629914 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-swiftconf\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.629979 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-ring-data-devices\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.630034 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-dispersionconf\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.630177 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5nzb\" (UniqueName: \"kubernetes.io/projected/996bb757-370f-4635-9b6f-923ecd70a8ee-kube-api-access-w5nzb\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.630295 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-combined-ca-bundle\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.630338 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-scripts\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.630395 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/996bb757-370f-4635-9b6f-923ecd70a8ee-etc-swift\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732255 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/996bb757-370f-4635-9b6f-923ecd70a8ee-etc-swift\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-swiftconf\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732356 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-ring-data-devices\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732378 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-dispersionconf\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732426 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5nzb\" (UniqueName: \"kubernetes.io/projected/996bb757-370f-4635-9b6f-923ecd70a8ee-kube-api-access-w5nzb\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732457 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-combined-ca-bundle\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.732492 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-scripts\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.733728 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-scripts\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.733924 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-ring-data-devices\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.734040 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/996bb757-370f-4635-9b6f-923ecd70a8ee-etc-swift\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.736396 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-dispersionconf\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.738062 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-swiftconf\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.738696 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-combined-ca-bundle\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.740890 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" event={"ID":"de4c5c56-929c-47b1-a488-f9eea7edeccc","Type":"ContainerStarted","Data":"8993fe3de7909e5d68edd6aa502520fae46a0622c78d34b6af3405418aee427a"} Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.749283 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5nzb\" (UniqueName: \"kubernetes.io/projected/996bb757-370f-4635-9b6f-923ecd70a8ee-kube-api-access-w5nzb\") pod \"swift-ring-rebalance-9bq7c\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.872926 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:28:45 crc kubenswrapper[4685]: I1204 06:28:45.935475 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:45 crc kubenswrapper[4685]: E1204 06:28:45.935798 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 06:28:45 crc kubenswrapper[4685]: E1204 06:28:45.935848 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 06:28:45 crc kubenswrapper[4685]: E1204 06:28:45.935940 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift podName:af46303b-ea28-4ab3-acb7-56a16afa6afd nodeName:}" failed. No retries permitted until 2025-12-04 06:28:46.935910434 +0000 UTC m=+1064.124141249 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift") pod "swift-storage-0" (UID: "af46303b-ea28-4ab3-acb7-56a16afa6afd") : configmap "swift-ring-files" not found Dec 04 06:28:46 crc kubenswrapper[4685]: I1204 06:28:46.386125 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9bq7c"] Dec 04 06:28:46 crc kubenswrapper[4685]: W1204 06:28:46.399923 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod996bb757_370f_4635_9b6f_923ecd70a8ee.slice/crio-214dbce78acad826508f02f913d0ef5d5ae665c99cfd1ea3f5d00bb37939b6a8 WatchSource:0}: Error finding container 214dbce78acad826508f02f913d0ef5d5ae665c99cfd1ea3f5d00bb37939b6a8: Status 404 returned error can't find the container with id 214dbce78acad826508f02f913d0ef5d5ae665c99cfd1ea3f5d00bb37939b6a8 Dec 04 06:28:46 crc kubenswrapper[4685]: I1204 06:28:46.752057 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9bq7c" event={"ID":"996bb757-370f-4635-9b6f-923ecd70a8ee","Type":"ContainerStarted","Data":"214dbce78acad826508f02f913d0ef5d5ae665c99cfd1ea3f5d00bb37939b6a8"} Dec 04 06:28:46 crc kubenswrapper[4685]: I1204 06:28:46.959350 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:46 crc kubenswrapper[4685]: E1204 06:28:46.959641 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 06:28:46 crc kubenswrapper[4685]: E1204 06:28:46.959686 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 06:28:46 crc kubenswrapper[4685]: E1204 06:28:46.959787 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift podName:af46303b-ea28-4ab3-acb7-56a16afa6afd nodeName:}" failed. No retries permitted until 2025-12-04 06:28:48.959755515 +0000 UTC m=+1066.147986320 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift") pod "swift-storage-0" (UID: "af46303b-ea28-4ab3-acb7-56a16afa6afd") : configmap "swift-ring-files" not found Dec 04 06:28:47 crc kubenswrapper[4685]: I1204 06:28:47.766169 4685 generic.go:334] "Generic (PLEG): container finished" podID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerID="251b7c3911050af23cf8810e5cd4c3046a1a62b7b20c47a86a29e0c37283ccfe" exitCode=0 Dec 04 06:28:47 crc kubenswrapper[4685]: I1204 06:28:47.766435 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" event={"ID":"de4c5c56-929c-47b1-a488-f9eea7edeccc","Type":"ContainerDied","Data":"251b7c3911050af23cf8810e5cd4c3046a1a62b7b20c47a86a29e0c37283ccfe"} Dec 04 06:28:47 crc kubenswrapper[4685]: I1204 06:28:47.773069 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-czlv2" event={"ID":"313ed371-bc49-447b-b609-7ec6a346e3f6","Type":"ContainerStarted","Data":"b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921"} Dec 04 06:28:47 crc kubenswrapper[4685]: I1204 06:28:47.773246 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:47 crc kubenswrapper[4685]: I1204 06:28:47.773228 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-czlv2" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerName="dnsmasq-dns" containerID="cri-o://b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921" gracePeriod=10 Dec 04 06:28:47 crc kubenswrapper[4685]: I1204 06:28:47.827435 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-czlv2" podStartSLOduration=7.827402397 podStartE2EDuration="7.827402397s" podCreationTimestamp="2025-12-04 06:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:28:47.823776579 +0000 UTC m=+1065.012007354" watchObservedRunningTime="2025-12-04 06:28:47.827402397 +0000 UTC m=+1065.015633172" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.382354 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.488664 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-dns-svc\") pod \"313ed371-bc49-447b-b609-7ec6a346e3f6\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.489039 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grtzt\" (UniqueName: \"kubernetes.io/projected/313ed371-bc49-447b-b609-7ec6a346e3f6-kube-api-access-grtzt\") pod \"313ed371-bc49-447b-b609-7ec6a346e3f6\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.489130 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-sb\") pod \"313ed371-bc49-447b-b609-7ec6a346e3f6\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.489188 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-config\") pod \"313ed371-bc49-447b-b609-7ec6a346e3f6\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.489222 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-nb\") pod \"313ed371-bc49-447b-b609-7ec6a346e3f6\" (UID: \"313ed371-bc49-447b-b609-7ec6a346e3f6\") " Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.496295 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/313ed371-bc49-447b-b609-7ec6a346e3f6-kube-api-access-grtzt" (OuterVolumeSpecName: "kube-api-access-grtzt") pod "313ed371-bc49-447b-b609-7ec6a346e3f6" (UID: "313ed371-bc49-447b-b609-7ec6a346e3f6"). InnerVolumeSpecName "kube-api-access-grtzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.540478 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "313ed371-bc49-447b-b609-7ec6a346e3f6" (UID: "313ed371-bc49-447b-b609-7ec6a346e3f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.542466 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "313ed371-bc49-447b-b609-7ec6a346e3f6" (UID: "313ed371-bc49-447b-b609-7ec6a346e3f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.548024 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "313ed371-bc49-447b-b609-7ec6a346e3f6" (UID: "313ed371-bc49-447b-b609-7ec6a346e3f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.548277 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-config" (OuterVolumeSpecName: "config") pod "313ed371-bc49-447b-b609-7ec6a346e3f6" (UID: "313ed371-bc49-447b-b609-7ec6a346e3f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.591005 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.591045 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grtzt\" (UniqueName: \"kubernetes.io/projected/313ed371-bc49-447b-b609-7ec6a346e3f6-kube-api-access-grtzt\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.591058 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.591070 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.591079 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/313ed371-bc49-447b-b609-7ec6a346e3f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.783987 4685 generic.go:334] "Generic (PLEG): container finished" podID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerID="b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921" exitCode=0 Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.784073 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-czlv2" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.784066 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-czlv2" event={"ID":"313ed371-bc49-447b-b609-7ec6a346e3f6","Type":"ContainerDied","Data":"b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921"} Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.784184 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-czlv2" event={"ID":"313ed371-bc49-447b-b609-7ec6a346e3f6","Type":"ContainerDied","Data":"ff0cc635cf11533eeb116e2abeb2e7ea8c2856801c827b6f693e5c9fdfd9d066"} Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.784231 4685 scope.go:117] "RemoveContainer" containerID="b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.787136 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" event={"ID":"de4c5c56-929c-47b1-a488-f9eea7edeccc","Type":"ContainerStarted","Data":"9dd67008de7ea673e410a7b4e0869ffe00a03de06e4557b6f96c2f152d815e7c"} Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.787527 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.789450 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"53bf9890-b3a9-4102-9b2f-f7711e7dbe70","Type":"ContainerStarted","Data":"dc720178a23973cd741daa09ef19b43b93aa39496f0c44ad8139d296c20d8b0d"} Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.789481 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"53bf9890-b3a9-4102-9b2f-f7711e7dbe70","Type":"ContainerStarted","Data":"78647c1900c53576bfd153f5bd8fe88ae796eeb526ce67f307f33604a01b3c26"} Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.789707 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.815284 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" podStartSLOduration=5.815260696 podStartE2EDuration="5.815260696s" podCreationTimestamp="2025-12-04 06:28:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:28:48.812322578 +0000 UTC m=+1066.000553353" watchObservedRunningTime="2025-12-04 06:28:48.815260696 +0000 UTC m=+1066.003491471" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.837224 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-czlv2"] Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.853093 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-czlv2"] Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.856888 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.890503443 podStartE2EDuration="8.856857193s" podCreationTimestamp="2025-12-04 06:28:40 +0000 UTC" firstStartedPulling="2025-12-04 06:28:41.19022693 +0000 UTC m=+1058.378457705" lastFinishedPulling="2025-12-04 06:28:48.15658068 +0000 UTC m=+1065.344811455" observedRunningTime="2025-12-04 06:28:48.84732897 +0000 UTC m=+1066.035559745" watchObservedRunningTime="2025-12-04 06:28:48.856857193 +0000 UTC m=+1066.045087968" Dec 04 06:28:48 crc kubenswrapper[4685]: I1204 06:28:48.997367 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:48 crc kubenswrapper[4685]: E1204 06:28:48.997646 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 06:28:48 crc kubenswrapper[4685]: E1204 06:28:48.997794 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 06:28:48 crc kubenswrapper[4685]: E1204 06:28:48.997873 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift podName:af46303b-ea28-4ab3-acb7-56a16afa6afd nodeName:}" failed. No retries permitted until 2025-12-04 06:28:52.997847628 +0000 UTC m=+1070.186078443 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift") pod "swift-storage-0" (UID: "af46303b-ea28-4ab3-acb7-56a16afa6afd") : configmap "swift-ring-files" not found Dec 04 06:28:49 crc kubenswrapper[4685]: I1204 06:28:49.143843 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" path="/var/lib/kubelet/pods/313ed371-bc49-447b-b609-7ec6a346e3f6/volumes" Dec 04 06:28:49 crc kubenswrapper[4685]: I1204 06:28:49.742396 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 06:28:49 crc kubenswrapper[4685]: I1204 06:28:49.832687 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 06:28:49 crc kubenswrapper[4685]: I1204 06:28:49.889093 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:49 crc kubenswrapper[4685]: I1204 06:28:49.974205 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 06:28:50 crc kubenswrapper[4685]: I1204 06:28:50.712718 4685 scope.go:117] "RemoveContainer" containerID="335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb" Dec 04 06:28:50 crc kubenswrapper[4685]: I1204 06:28:50.779373 4685 scope.go:117] "RemoveContainer" containerID="b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921" Dec 04 06:28:50 crc kubenswrapper[4685]: E1204 06:28:50.780022 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921\": container with ID starting with b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921 not found: ID does not exist" containerID="b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921" Dec 04 06:28:50 crc kubenswrapper[4685]: I1204 06:28:50.780065 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921"} err="failed to get container status \"b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921\": rpc error: code = NotFound desc = could not find container \"b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921\": container with ID starting with b440e6ceedd3ead46ae4de9721978c606f957cd2b9bbab870fda05fdc087f921 not found: ID does not exist" Dec 04 06:28:50 crc kubenswrapper[4685]: I1204 06:28:50.780115 4685 scope.go:117] "RemoveContainer" containerID="335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb" Dec 04 06:28:50 crc kubenswrapper[4685]: E1204 06:28:50.780423 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb\": container with ID starting with 335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb not found: ID does not exist" containerID="335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb" Dec 04 06:28:50 crc kubenswrapper[4685]: I1204 06:28:50.780448 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb"} err="failed to get container status \"335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb\": rpc error: code = NotFound desc = could not find container \"335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb\": container with ID starting with 335380fb6db2255a663c60a24bd2fdde90e9eccffa02efbdb81e798745efcffb not found: ID does not exist" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.233370 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jn7p4"] Dec 04 06:28:51 crc kubenswrapper[4685]: E1204 06:28:51.234364 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerName="init" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.234381 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerName="init" Dec 04 06:28:51 crc kubenswrapper[4685]: E1204 06:28:51.234435 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerName="dnsmasq-dns" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.234443 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerName="dnsmasq-dns" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.234635 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="313ed371-bc49-447b-b609-7ec6a346e3f6" containerName="dnsmasq-dns" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.235337 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.243290 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jn7p4"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.324997 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cd77-account-create-update-2vkcr"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.326696 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.329616 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.342574 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-operator-scripts\") pod \"keystone-db-create-jn7p4\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.342756 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5wfs\" (UniqueName: \"kubernetes.io/projected/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-kube-api-access-p5wfs\") pod \"keystone-db-create-jn7p4\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.348277 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cd77-account-create-update-2vkcr"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.416124 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-kfvw5"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.417790 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.426299 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kfvw5"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.444449 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-operator-scripts\") pod \"keystone-db-create-jn7p4\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.444562 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn2rm\" (UniqueName: \"kubernetes.io/projected/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-kube-api-access-sn2rm\") pod \"keystone-cd77-account-create-update-2vkcr\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.444591 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5wfs\" (UniqueName: \"kubernetes.io/projected/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-kube-api-access-p5wfs\") pod \"keystone-db-create-jn7p4\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.444644 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-operator-scripts\") pod \"keystone-cd77-account-create-update-2vkcr\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.445432 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-operator-scripts\") pod \"keystone-db-create-jn7p4\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.463286 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5wfs\" (UniqueName: \"kubernetes.io/projected/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-kube-api-access-p5wfs\") pod \"keystone-db-create-jn7p4\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.533332 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8a6e-account-create-update-vhvkz"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.534499 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.536836 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.546092 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cf58941-8b22-419f-bcd6-27bd436ddde4-operator-scripts\") pod \"placement-db-create-kfvw5\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.547068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn2rm\" (UniqueName: \"kubernetes.io/projected/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-kube-api-access-sn2rm\") pod \"keystone-cd77-account-create-update-2vkcr\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.547629 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75ksp\" (UniqueName: \"kubernetes.io/projected/4cf58941-8b22-419f-bcd6-27bd436ddde4-kube-api-access-75ksp\") pod \"placement-db-create-kfvw5\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.547942 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-operator-scripts\") pod \"keystone-cd77-account-create-update-2vkcr\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.548928 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-operator-scripts\") pod \"keystone-cd77-account-create-update-2vkcr\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.551761 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8a6e-account-create-update-vhvkz"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.556278 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.571531 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn2rm\" (UniqueName: \"kubernetes.io/projected/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-kube-api-access-sn2rm\") pod \"keystone-cd77-account-create-update-2vkcr\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.648363 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.664719 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cf58941-8b22-419f-bcd6-27bd436ddde4-operator-scripts\") pod \"placement-db-create-kfvw5\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.664999 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6edc5c45-5581-4a04-9414-847a4c817439-operator-scripts\") pod \"placement-8a6e-account-create-update-vhvkz\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.665066 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75ksp\" (UniqueName: \"kubernetes.io/projected/4cf58941-8b22-419f-bcd6-27bd436ddde4-kube-api-access-75ksp\") pod \"placement-db-create-kfvw5\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.665120 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpdwl\" (UniqueName: \"kubernetes.io/projected/6edc5c45-5581-4a04-9414-847a4c817439-kube-api-access-gpdwl\") pod \"placement-8a6e-account-create-update-vhvkz\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.666725 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cf58941-8b22-419f-bcd6-27bd436ddde4-operator-scripts\") pod \"placement-db-create-kfvw5\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.688738 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75ksp\" (UniqueName: \"kubernetes.io/projected/4cf58941-8b22-419f-bcd6-27bd436ddde4-kube-api-access-75ksp\") pod \"placement-db-create-kfvw5\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.729541 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dsnhj"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.730746 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.741451 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.742184 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dsnhj"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.767486 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6edc5c45-5581-4a04-9414-847a4c817439-operator-scripts\") pod \"placement-8a6e-account-create-update-vhvkz\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.768081 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpdwl\" (UniqueName: \"kubernetes.io/projected/6edc5c45-5581-4a04-9414-847a4c817439-kube-api-access-gpdwl\") pod \"placement-8a6e-account-create-update-vhvkz\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.768623 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6edc5c45-5581-4a04-9414-847a4c817439-operator-scripts\") pod \"placement-8a6e-account-create-update-vhvkz\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.788625 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpdwl\" (UniqueName: \"kubernetes.io/projected/6edc5c45-5581-4a04-9414-847a4c817439-kube-api-access-gpdwl\") pod \"placement-8a6e-account-create-update-vhvkz\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.853780 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.854300 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9bq7c" event={"ID":"996bb757-370f-4635-9b6f-923ecd70a8ee","Type":"ContainerStarted","Data":"122485dce8a8a6632de00e9ea10d9bfd51891c3f56a0528b82f9bedd31a5275d"} Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.869867 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-operator-scripts\") pod \"glance-db-create-dsnhj\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.870063 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn5tr\" (UniqueName: \"kubernetes.io/projected/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-kube-api-access-nn5tr\") pod \"glance-db-create-dsnhj\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.923364 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9bq7c" podStartSLOduration=2.5137937470000002 podStartE2EDuration="6.923335491s" podCreationTimestamp="2025-12-04 06:28:45 +0000 UTC" firstStartedPulling="2025-12-04 06:28:46.404528796 +0000 UTC m=+1063.592759571" lastFinishedPulling="2025-12-04 06:28:50.81407052 +0000 UTC m=+1068.002301315" observedRunningTime="2025-12-04 06:28:51.875719745 +0000 UTC m=+1069.063950520" watchObservedRunningTime="2025-12-04 06:28:51.923335491 +0000 UTC m=+1069.111566266" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.932472 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9056-account-create-update-2sqd8"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.960728 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.968925 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.976045 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-operator-scripts\") pod \"glance-db-create-dsnhj\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.976324 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn5tr\" (UniqueName: \"kubernetes.io/projected/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-kube-api-access-nn5tr\") pod \"glance-db-create-dsnhj\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.979374 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-operator-scripts\") pod \"glance-db-create-dsnhj\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.990814 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9056-account-create-update-2sqd8"] Dec 04 06:28:51 crc kubenswrapper[4685]: I1204 06:28:51.998589 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn5tr\" (UniqueName: \"kubernetes.io/projected/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-kube-api-access-nn5tr\") pod \"glance-db-create-dsnhj\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.060633 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jn7p4"] Dec 04 06:28:52 crc kubenswrapper[4685]: W1204 06:28:52.062545 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9214ce67_d976_4159_9a3b_dcc8d5ff3c1d.slice/crio-52ca02294892eb507e84766863889383129ffd8269672b428f497c03701c4cbd WatchSource:0}: Error finding container 52ca02294892eb507e84766863889383129ffd8269672b428f497c03701c4cbd: Status 404 returned error can't find the container with id 52ca02294892eb507e84766863889383129ffd8269672b428f497c03701c4cbd Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.078690 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78799f6c-2669-4a97-b50d-a772ee9b7dea-operator-scripts\") pod \"glance-9056-account-create-update-2sqd8\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.078780 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2qg2\" (UniqueName: \"kubernetes.io/projected/78799f6c-2669-4a97-b50d-a772ee9b7dea-kube-api-access-h2qg2\") pod \"glance-9056-account-create-update-2sqd8\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.130015 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.180791 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78799f6c-2669-4a97-b50d-a772ee9b7dea-operator-scripts\") pod \"glance-9056-account-create-update-2sqd8\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.180857 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2qg2\" (UniqueName: \"kubernetes.io/projected/78799f6c-2669-4a97-b50d-a772ee9b7dea-kube-api-access-h2qg2\") pod \"glance-9056-account-create-update-2sqd8\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.194134 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78799f6c-2669-4a97-b50d-a772ee9b7dea-operator-scripts\") pod \"glance-9056-account-create-update-2sqd8\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.199113 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cd77-account-create-update-2vkcr"] Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.211587 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2qg2\" (UniqueName: \"kubernetes.io/projected/78799f6c-2669-4a97-b50d-a772ee9b7dea-kube-api-access-h2qg2\") pod \"glance-9056-account-create-update-2sqd8\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.292587 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.369947 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kfvw5"] Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.459038 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8a6e-account-create-update-vhvkz"] Dec 04 06:28:52 crc kubenswrapper[4685]: W1204 06:28:52.470445 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6edc5c45_5581_4a04_9414_847a4c817439.slice/crio-66373df2527c78f25a949356c82ba626716853e601a433bcb3072b1eff85cff6 WatchSource:0}: Error finding container 66373df2527c78f25a949356c82ba626716853e601a433bcb3072b1eff85cff6: Status 404 returned error can't find the container with id 66373df2527c78f25a949356c82ba626716853e601a433bcb3072b1eff85cff6 Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.588138 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dsnhj"] Dec 04 06:28:52 crc kubenswrapper[4685]: W1204 06:28:52.670763 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab4e10f7_15cc_42da_9a4b_80738fb81cb2.slice/crio-5bad7bd61d48c001c5337ebb81928b5caea43e675e37167818c03690ba0d6ee2 WatchSource:0}: Error finding container 5bad7bd61d48c001c5337ebb81928b5caea43e675e37167818c03690ba0d6ee2: Status 404 returned error can't find the container with id 5bad7bd61d48c001c5337ebb81928b5caea43e675e37167818c03690ba0d6ee2 Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.744483 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9056-account-create-update-2sqd8"] Dec 04 06:28:52 crc kubenswrapper[4685]: W1204 06:28:52.827017 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78799f6c_2669_4a97_b50d_a772ee9b7dea.slice/crio-9a91e7be19a187dc7438f3517011a63c78ba760f80ec6d905b4d21a1db28c64d WatchSource:0}: Error finding container 9a91e7be19a187dc7438f3517011a63c78ba760f80ec6d905b4d21a1db28c64d: Status 404 returned error can't find the container with id 9a91e7be19a187dc7438f3517011a63c78ba760f80ec6d905b4d21a1db28c64d Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.876924 4685 generic.go:334] "Generic (PLEG): container finished" podID="e844c8fb-db9a-4bc4-a0a1-4fe116f05608" containerID="d50bbb1bb2816fc757797cccc70048dd6e28dd963f4e1b16296b8ed76e964693" exitCode=0 Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.877007 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd77-account-create-update-2vkcr" event={"ID":"e844c8fb-db9a-4bc4-a0a1-4fe116f05608","Type":"ContainerDied","Data":"d50bbb1bb2816fc757797cccc70048dd6e28dd963f4e1b16296b8ed76e964693"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.877046 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd77-account-create-update-2vkcr" event={"ID":"e844c8fb-db9a-4bc4-a0a1-4fe116f05608","Type":"ContainerStarted","Data":"07f6e94258ad6f06f794a488f32512a585a972e4c00c57a546d66018fdbb2d6c"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.884720 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a6e-account-create-update-vhvkz" event={"ID":"6edc5c45-5581-4a04-9414-847a4c817439","Type":"ContainerStarted","Data":"dba855c2804e36e3af395966a469b7b9eac03d6d2f0243dcc8253af7ed252f1d"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.884770 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a6e-account-create-update-vhvkz" event={"ID":"6edc5c45-5581-4a04-9414-847a4c817439","Type":"ContainerStarted","Data":"66373df2527c78f25a949356c82ba626716853e601a433bcb3072b1eff85cff6"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.893057 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9056-account-create-update-2sqd8" event={"ID":"78799f6c-2669-4a97-b50d-a772ee9b7dea","Type":"ContainerStarted","Data":"9a91e7be19a187dc7438f3517011a63c78ba760f80ec6d905b4d21a1db28c64d"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.895592 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dsnhj" event={"ID":"ab4e10f7-15cc-42da-9a4b-80738fb81cb2","Type":"ContainerStarted","Data":"5bad7bd61d48c001c5337ebb81928b5caea43e675e37167818c03690ba0d6ee2"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.900692 4685 generic.go:334] "Generic (PLEG): container finished" podID="4cf58941-8b22-419f-bcd6-27bd436ddde4" containerID="7f80d8af48f1348c02c77eb50dbdf263cd1ace923a39ca6358f662f61a1a4cb4" exitCode=0 Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.900801 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kfvw5" event={"ID":"4cf58941-8b22-419f-bcd6-27bd436ddde4","Type":"ContainerDied","Data":"7f80d8af48f1348c02c77eb50dbdf263cd1ace923a39ca6358f662f61a1a4cb4"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.900837 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kfvw5" event={"ID":"4cf58941-8b22-419f-bcd6-27bd436ddde4","Type":"ContainerStarted","Data":"153d5e342ba9d597a64e964faded2475dc43c61fa09026e7eca7ca09ff9a9643"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.904916 4685 generic.go:334] "Generic (PLEG): container finished" podID="9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" containerID="3dc417c7088031ac631b07ca8f46de81e93ee3587153b26c3f906708394d9e5b" exitCode=0 Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.905138 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jn7p4" event={"ID":"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d","Type":"ContainerDied","Data":"3dc417c7088031ac631b07ca8f46de81e93ee3587153b26c3f906708394d9e5b"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.905183 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jn7p4" event={"ID":"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d","Type":"ContainerStarted","Data":"52ca02294892eb507e84766863889383129ffd8269672b428f497c03701c4cbd"} Dec 04 06:28:52 crc kubenswrapper[4685]: I1204 06:28:52.944090 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8a6e-account-create-update-vhvkz" podStartSLOduration=1.944067948 podStartE2EDuration="1.944067948s" podCreationTimestamp="2025-12-04 06:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:28:52.937961926 +0000 UTC m=+1070.126192701" watchObservedRunningTime="2025-12-04 06:28:52.944067948 +0000 UTC m=+1070.132298723" Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.097126 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:28:53 crc kubenswrapper[4685]: E1204 06:28:53.097458 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 06:28:53 crc kubenswrapper[4685]: E1204 06:28:53.097506 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 06:28:53 crc kubenswrapper[4685]: E1204 06:28:53.097599 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift podName:af46303b-ea28-4ab3-acb7-56a16afa6afd nodeName:}" failed. No retries permitted until 2025-12-04 06:29:01.097573815 +0000 UTC m=+1078.285804600 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift") pod "swift-storage-0" (UID: "af46303b-ea28-4ab3-acb7-56a16afa6afd") : configmap "swift-ring-files" not found Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.918502 4685 generic.go:334] "Generic (PLEG): container finished" podID="6edc5c45-5581-4a04-9414-847a4c817439" containerID="dba855c2804e36e3af395966a469b7b9eac03d6d2f0243dcc8253af7ed252f1d" exitCode=0 Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.918673 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a6e-account-create-update-vhvkz" event={"ID":"6edc5c45-5581-4a04-9414-847a4c817439","Type":"ContainerDied","Data":"dba855c2804e36e3af395966a469b7b9eac03d6d2f0243dcc8253af7ed252f1d"} Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.921032 4685 generic.go:334] "Generic (PLEG): container finished" podID="78799f6c-2669-4a97-b50d-a772ee9b7dea" containerID="c43e0b227c96c5ec640cb727d5443cab2edb0ee417eaffa427af003de0a8f88f" exitCode=0 Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.921090 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9056-account-create-update-2sqd8" event={"ID":"78799f6c-2669-4a97-b50d-a772ee9b7dea","Type":"ContainerDied","Data":"c43e0b227c96c5ec640cb727d5443cab2edb0ee417eaffa427af003de0a8f88f"} Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.923084 4685 generic.go:334] "Generic (PLEG): container finished" podID="ab4e10f7-15cc-42da-9a4b-80738fb81cb2" containerID="e67ca61666265757fdc38ab13f318dea4d38ca0f1be22bc8338befdeec1a7deb" exitCode=0 Dec 04 06:28:53 crc kubenswrapper[4685]: I1204 06:28:53.923307 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dsnhj" event={"ID":"ab4e10f7-15cc-42da-9a4b-80738fb81cb2","Type":"ContainerDied","Data":"e67ca61666265757fdc38ab13f318dea4d38ca0f1be22bc8338befdeec1a7deb"} Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.347543 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.463521 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6wc8c"] Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.463932 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerName="dnsmasq-dns" containerID="cri-o://1fd45d68951d01729b841414b884c7a72b5937aa612b095d00f9b4c1a4711998" gracePeriod=10 Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.602654 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.611051 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.617819 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.737501 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-operator-scripts\") pod \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.737643 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn2rm\" (UniqueName: \"kubernetes.io/projected/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-kube-api-access-sn2rm\") pod \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.737755 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75ksp\" (UniqueName: \"kubernetes.io/projected/4cf58941-8b22-419f-bcd6-27bd436ddde4-kube-api-access-75ksp\") pod \"4cf58941-8b22-419f-bcd6-27bd436ddde4\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.737798 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cf58941-8b22-419f-bcd6-27bd436ddde4-operator-scripts\") pod \"4cf58941-8b22-419f-bcd6-27bd436ddde4\" (UID: \"4cf58941-8b22-419f-bcd6-27bd436ddde4\") " Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.737877 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-operator-scripts\") pod \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\" (UID: \"e844c8fb-db9a-4bc4-a0a1-4fe116f05608\") " Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.738001 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5wfs\" (UniqueName: \"kubernetes.io/projected/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-kube-api-access-p5wfs\") pod \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\" (UID: \"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d\") " Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.738869 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf58941-8b22-419f-bcd6-27bd436ddde4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4cf58941-8b22-419f-bcd6-27bd436ddde4" (UID: "4cf58941-8b22-419f-bcd6-27bd436ddde4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.738919 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" (UID: "9214ce67-d976-4159-9a3b-dcc8d5ff3c1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.739445 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e844c8fb-db9a-4bc4-a0a1-4fe116f05608" (UID: "e844c8fb-db9a-4bc4-a0a1-4fe116f05608"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.745479 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-kube-api-access-sn2rm" (OuterVolumeSpecName: "kube-api-access-sn2rm") pod "e844c8fb-db9a-4bc4-a0a1-4fe116f05608" (UID: "e844c8fb-db9a-4bc4-a0a1-4fe116f05608"). InnerVolumeSpecName "kube-api-access-sn2rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.745677 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf58941-8b22-419f-bcd6-27bd436ddde4-kube-api-access-75ksp" (OuterVolumeSpecName: "kube-api-access-75ksp") pod "4cf58941-8b22-419f-bcd6-27bd436ddde4" (UID: "4cf58941-8b22-419f-bcd6-27bd436ddde4"). InnerVolumeSpecName "kube-api-access-75ksp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.746498 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-kube-api-access-p5wfs" (OuterVolumeSpecName: "kube-api-access-p5wfs") pod "9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" (UID: "9214ce67-d976-4159-9a3b-dcc8d5ff3c1d"). InnerVolumeSpecName "kube-api-access-p5wfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.840231 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5wfs\" (UniqueName: \"kubernetes.io/projected/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-kube-api-access-p5wfs\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.840553 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.840568 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn2rm\" (UniqueName: \"kubernetes.io/projected/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-kube-api-access-sn2rm\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.840582 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75ksp\" (UniqueName: \"kubernetes.io/projected/4cf58941-8b22-419f-bcd6-27bd436ddde4-kube-api-access-75ksp\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.840594 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cf58941-8b22-419f-bcd6-27bd436ddde4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.840603 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e844c8fb-db9a-4bc4-a0a1-4fe116f05608-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.943192 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd77-account-create-update-2vkcr" event={"ID":"e844c8fb-db9a-4bc4-a0a1-4fe116f05608","Type":"ContainerDied","Data":"07f6e94258ad6f06f794a488f32512a585a972e4c00c57a546d66018fdbb2d6c"} Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.943248 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07f6e94258ad6f06f794a488f32512a585a972e4c00c57a546d66018fdbb2d6c" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.943324 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd77-account-create-update-2vkcr" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.952827 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kfvw5" event={"ID":"4cf58941-8b22-419f-bcd6-27bd436ddde4","Type":"ContainerDied","Data":"153d5e342ba9d597a64e964faded2475dc43c61fa09026e7eca7ca09ff9a9643"} Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.952889 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="153d5e342ba9d597a64e964faded2475dc43c61fa09026e7eca7ca09ff9a9643" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.952973 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kfvw5" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.955091 4685 generic.go:334] "Generic (PLEG): container finished" podID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerID="1fd45d68951d01729b841414b884c7a72b5937aa612b095d00f9b4c1a4711998" exitCode=0 Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.955155 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" event={"ID":"6f031df1-b162-4ace-8c47-021a7d1c9e61","Type":"ContainerDied","Data":"1fd45d68951d01729b841414b884c7a72b5937aa612b095d00f9b4c1a4711998"} Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.956727 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jn7p4" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.959307 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jn7p4" event={"ID":"9214ce67-d976-4159-9a3b-dcc8d5ff3c1d","Type":"ContainerDied","Data":"52ca02294892eb507e84766863889383129ffd8269672b428f497c03701c4cbd"} Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.959462 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52ca02294892eb507e84766863889383129ffd8269672b428f497c03701c4cbd" Dec 04 06:28:54 crc kubenswrapper[4685]: I1204 06:28:54.993623 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.045495 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lkcc\" (UniqueName: \"kubernetes.io/projected/6f031df1-b162-4ace-8c47-021a7d1c9e61-kube-api-access-5lkcc\") pod \"6f031df1-b162-4ace-8c47-021a7d1c9e61\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.045740 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-dns-svc\") pod \"6f031df1-b162-4ace-8c47-021a7d1c9e61\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.045831 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-config\") pod \"6f031df1-b162-4ace-8c47-021a7d1c9e61\" (UID: \"6f031df1-b162-4ace-8c47-021a7d1c9e61\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.053704 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f031df1-b162-4ace-8c47-021a7d1c9e61-kube-api-access-5lkcc" (OuterVolumeSpecName: "kube-api-access-5lkcc") pod "6f031df1-b162-4ace-8c47-021a7d1c9e61" (UID: "6f031df1-b162-4ace-8c47-021a7d1c9e61"). InnerVolumeSpecName "kube-api-access-5lkcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.101849 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-config" (OuterVolumeSpecName: "config") pod "6f031df1-b162-4ace-8c47-021a7d1c9e61" (UID: "6f031df1-b162-4ace-8c47-021a7d1c9e61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.106797 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f031df1-b162-4ace-8c47-021a7d1c9e61" (UID: "6f031df1-b162-4ace-8c47-021a7d1c9e61"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.148353 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.148389 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f031df1-b162-4ace-8c47-021a7d1c9e61-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.148404 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lkcc\" (UniqueName: \"kubernetes.io/projected/6f031df1-b162-4ace-8c47-021a7d1c9e61-kube-api-access-5lkcc\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.561506 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.565080 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.573794 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.661255 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpdwl\" (UniqueName: \"kubernetes.io/projected/6edc5c45-5581-4a04-9414-847a4c817439-kube-api-access-gpdwl\") pod \"6edc5c45-5581-4a04-9414-847a4c817439\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.661375 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2qg2\" (UniqueName: \"kubernetes.io/projected/78799f6c-2669-4a97-b50d-a772ee9b7dea-kube-api-access-h2qg2\") pod \"78799f6c-2669-4a97-b50d-a772ee9b7dea\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.661428 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6edc5c45-5581-4a04-9414-847a4c817439-operator-scripts\") pod \"6edc5c45-5581-4a04-9414-847a4c817439\" (UID: \"6edc5c45-5581-4a04-9414-847a4c817439\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.661470 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-operator-scripts\") pod \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.661506 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78799f6c-2669-4a97-b50d-a772ee9b7dea-operator-scripts\") pod \"78799f6c-2669-4a97-b50d-a772ee9b7dea\" (UID: \"78799f6c-2669-4a97-b50d-a772ee9b7dea\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.661598 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn5tr\" (UniqueName: \"kubernetes.io/projected/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-kube-api-access-nn5tr\") pod \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\" (UID: \"ab4e10f7-15cc-42da-9a4b-80738fb81cb2\") " Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.663668 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edc5c45-5581-4a04-9414-847a4c817439-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6edc5c45-5581-4a04-9414-847a4c817439" (UID: "6edc5c45-5581-4a04-9414-847a4c817439"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.664785 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab4e10f7-15cc-42da-9a4b-80738fb81cb2" (UID: "ab4e10f7-15cc-42da-9a4b-80738fb81cb2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.665716 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78799f6c-2669-4a97-b50d-a772ee9b7dea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "78799f6c-2669-4a97-b50d-a772ee9b7dea" (UID: "78799f6c-2669-4a97-b50d-a772ee9b7dea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.666653 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-kube-api-access-nn5tr" (OuterVolumeSpecName: "kube-api-access-nn5tr") pod "ab4e10f7-15cc-42da-9a4b-80738fb81cb2" (UID: "ab4e10f7-15cc-42da-9a4b-80738fb81cb2"). InnerVolumeSpecName "kube-api-access-nn5tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.669482 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78799f6c-2669-4a97-b50d-a772ee9b7dea-kube-api-access-h2qg2" (OuterVolumeSpecName: "kube-api-access-h2qg2") pod "78799f6c-2669-4a97-b50d-a772ee9b7dea" (UID: "78799f6c-2669-4a97-b50d-a772ee9b7dea"). InnerVolumeSpecName "kube-api-access-h2qg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.674662 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6edc5c45-5581-4a04-9414-847a4c817439-kube-api-access-gpdwl" (OuterVolumeSpecName: "kube-api-access-gpdwl") pod "6edc5c45-5581-4a04-9414-847a4c817439" (UID: "6edc5c45-5581-4a04-9414-847a4c817439"). InnerVolumeSpecName "kube-api-access-gpdwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.763801 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.764151 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78799f6c-2669-4a97-b50d-a772ee9b7dea-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.764235 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn5tr\" (UniqueName: \"kubernetes.io/projected/ab4e10f7-15cc-42da-9a4b-80738fb81cb2-kube-api-access-nn5tr\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.764303 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpdwl\" (UniqueName: \"kubernetes.io/projected/6edc5c45-5581-4a04-9414-847a4c817439-kube-api-access-gpdwl\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.764373 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2qg2\" (UniqueName: \"kubernetes.io/projected/78799f6c-2669-4a97-b50d-a772ee9b7dea-kube-api-access-h2qg2\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.764455 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6edc5c45-5581-4a04-9414-847a4c817439-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.967753 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" event={"ID":"6f031df1-b162-4ace-8c47-021a7d1c9e61","Type":"ContainerDied","Data":"0f3c328a3a803b3b509841b3a7dfc4b1847d2a6da6c15fa8ede55d931ac465c4"} Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.968326 4685 scope.go:117] "RemoveContainer" containerID="1fd45d68951d01729b841414b884c7a72b5937aa612b095d00f9b4c1a4711998" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.967804 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6wc8c" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.972630 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a6e-account-create-update-vhvkz" event={"ID":"6edc5c45-5581-4a04-9414-847a4c817439","Type":"ContainerDied","Data":"66373df2527c78f25a949356c82ba626716853e601a433bcb3072b1eff85cff6"} Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.972663 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66373df2527c78f25a949356c82ba626716853e601a433bcb3072b1eff85cff6" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.973203 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a6e-account-create-update-vhvkz" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.974243 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9056-account-create-update-2sqd8" event={"ID":"78799f6c-2669-4a97-b50d-a772ee9b7dea","Type":"ContainerDied","Data":"9a91e7be19a187dc7438f3517011a63c78ba760f80ec6d905b4d21a1db28c64d"} Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.974280 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a91e7be19a187dc7438f3517011a63c78ba760f80ec6d905b4d21a1db28c64d" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.974365 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9056-account-create-update-2sqd8" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.976239 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dsnhj" event={"ID":"ab4e10f7-15cc-42da-9a4b-80738fb81cb2","Type":"ContainerDied","Data":"5bad7bd61d48c001c5337ebb81928b5caea43e675e37167818c03690ba0d6ee2"} Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.976267 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bad7bd61d48c001c5337ebb81928b5caea43e675e37167818c03690ba0d6ee2" Dec 04 06:28:55 crc kubenswrapper[4685]: I1204 06:28:55.976321 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dsnhj" Dec 04 06:28:56 crc kubenswrapper[4685]: I1204 06:28:56.016517 4685 scope.go:117] "RemoveContainer" containerID="3f7646ff4e72435b72c62993daf80bcfbd0f281b3d771c84f1c67022a2eaa3a9" Dec 04 06:28:56 crc kubenswrapper[4685]: I1204 06:28:56.018836 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6wc8c"] Dec 04 06:28:56 crc kubenswrapper[4685]: I1204 06:28:56.029423 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6wc8c"] Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.075198 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8bkmd"] Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.076590 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6edc5c45-5581-4a04-9414-847a4c817439" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.076701 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6edc5c45-5581-4a04-9414-847a4c817439" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.076777 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e844c8fb-db9a-4bc4-a0a1-4fe116f05608" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.076878 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e844c8fb-db9a-4bc4-a0a1-4fe116f05608" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.076932 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4e10f7-15cc-42da-9a4b-80738fb81cb2" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.076982 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4e10f7-15cc-42da-9a4b-80738fb81cb2" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.077049 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerName="init" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.077109 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerName="init" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.077167 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf58941-8b22-419f-bcd6-27bd436ddde4" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.077222 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf58941-8b22-419f-bcd6-27bd436ddde4" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.077283 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78799f6c-2669-4a97-b50d-a772ee9b7dea" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.077334 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="78799f6c-2669-4a97-b50d-a772ee9b7dea" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.077400 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerName="dnsmasq-dns" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.077488 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerName="dnsmasq-dns" Dec 04 06:28:57 crc kubenswrapper[4685]: E1204 06:28:57.077545 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.077606 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.077932 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4e10f7-15cc-42da-9a4b-80738fb81cb2" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.078029 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf58941-8b22-419f-bcd6-27bd436ddde4" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.078105 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="78799f6c-2669-4a97-b50d-a772ee9b7dea" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.078166 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" containerName="mariadb-database-create" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.078221 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6edc5c45-5581-4a04-9414-847a4c817439" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.078286 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" containerName="dnsmasq-dns" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.078353 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e844c8fb-db9a-4bc4-a0a1-4fe116f05608" containerName="mariadb-account-create-update" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.079298 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.082600 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8bkmd"] Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.083218 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-66ml8" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.084491 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.137253 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f031df1-b162-4ace-8c47-021a7d1c9e61" path="/var/lib/kubelet/pods/6f031df1-b162-4ace-8c47-021a7d1c9e61/volumes" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.203848 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz2jc\" (UniqueName: \"kubernetes.io/projected/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-kube-api-access-wz2jc\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.204150 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-combined-ca-bundle\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.204318 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-config-data\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.204463 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-db-sync-config-data\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.306307 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz2jc\" (UniqueName: \"kubernetes.io/projected/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-kube-api-access-wz2jc\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.306369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-combined-ca-bundle\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.306464 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-config-data\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.306506 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-db-sync-config-data\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.313536 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-config-data\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.314383 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-combined-ca-bundle\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.314623 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-db-sync-config-data\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.326894 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz2jc\" (UniqueName: \"kubernetes.io/projected/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-kube-api-access-wz2jc\") pod \"glance-db-sync-8bkmd\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:57 crc kubenswrapper[4685]: I1204 06:28:57.397831 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8bkmd" Dec 04 06:28:58 crc kubenswrapper[4685]: I1204 06:28:58.003318 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8bkmd"] Dec 04 06:28:58 crc kubenswrapper[4685]: W1204 06:28:58.005918 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2733d0b6_b52b_4e2e_836b_5dd09dc7b2f0.slice/crio-f68aab5abc5f4913f6afc3186cff505af58115c49e859e5a182b24b31ac7eb6c WatchSource:0}: Error finding container f68aab5abc5f4913f6afc3186cff505af58115c49e859e5a182b24b31ac7eb6c: Status 404 returned error can't find the container with id f68aab5abc5f4913f6afc3186cff505af58115c49e859e5a182b24b31ac7eb6c Dec 04 06:28:58 crc kubenswrapper[4685]: E1204 06:28:58.702609 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod996bb757_370f_4635_9b6f_923ecd70a8ee.slice/crio-122485dce8a8a6632de00e9ea10d9bfd51891c3f56a0528b82f9bedd31a5275d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod996bb757_370f_4635_9b6f_923ecd70a8ee.slice/crio-conmon-122485dce8a8a6632de00e9ea10d9bfd51891c3f56a0528b82f9bedd31a5275d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:28:59 crc kubenswrapper[4685]: I1204 06:28:59.009265 4685 generic.go:334] "Generic (PLEG): container finished" podID="996bb757-370f-4635-9b6f-923ecd70a8ee" containerID="122485dce8a8a6632de00e9ea10d9bfd51891c3f56a0528b82f9bedd31a5275d" exitCode=0 Dec 04 06:28:59 crc kubenswrapper[4685]: I1204 06:28:59.009373 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9bq7c" event={"ID":"996bb757-370f-4635-9b6f-923ecd70a8ee","Type":"ContainerDied","Data":"122485dce8a8a6632de00e9ea10d9bfd51891c3f56a0528b82f9bedd31a5275d"} Dec 04 06:28:59 crc kubenswrapper[4685]: I1204 06:28:59.012766 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8bkmd" event={"ID":"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0","Type":"ContainerStarted","Data":"f68aab5abc5f4913f6afc3186cff505af58115c49e859e5a182b24b31ac7eb6c"} Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.472856 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567197 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-combined-ca-bundle\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567316 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-swiftconf\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-scripts\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567570 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-ring-data-devices\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567612 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-dispersionconf\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567638 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5nzb\" (UniqueName: \"kubernetes.io/projected/996bb757-370f-4635-9b6f-923ecd70a8ee-kube-api-access-w5nzb\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.567671 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/996bb757-370f-4635-9b6f-923ecd70a8ee-etc-swift\") pod \"996bb757-370f-4635-9b6f-923ecd70a8ee\" (UID: \"996bb757-370f-4635-9b6f-923ecd70a8ee\") " Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.569037 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.569142 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/996bb757-370f-4635-9b6f-923ecd70a8ee-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.576362 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/996bb757-370f-4635-9b6f-923ecd70a8ee-kube-api-access-w5nzb" (OuterVolumeSpecName: "kube-api-access-w5nzb") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "kube-api-access-w5nzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.580431 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.593158 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-scripts" (OuterVolumeSpecName: "scripts") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.595786 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.596949 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "996bb757-370f-4635-9b6f-923ecd70a8ee" (UID: "996bb757-370f-4635-9b6f-923ecd70a8ee"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.668847 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669774 4685 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669792 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669802 4685 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/996bb757-370f-4635-9b6f-923ecd70a8ee-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669814 4685 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669824 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5nzb\" (UniqueName: \"kubernetes.io/projected/996bb757-370f-4635-9b6f-923ecd70a8ee-kube-api-access-w5nzb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669833 4685 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/996bb757-370f-4635-9b6f-923ecd70a8ee-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:00 crc kubenswrapper[4685]: I1204 06:29:00.669842 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996bb757-370f-4635-9b6f-923ecd70a8ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:01 crc kubenswrapper[4685]: I1204 06:29:01.049201 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9bq7c" event={"ID":"996bb757-370f-4635-9b6f-923ecd70a8ee","Type":"ContainerDied","Data":"214dbce78acad826508f02f913d0ef5d5ae665c99cfd1ea3f5d00bb37939b6a8"} Dec 04 06:29:01 crc kubenswrapper[4685]: I1204 06:29:01.049267 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="214dbce78acad826508f02f913d0ef5d5ae665c99cfd1ea3f5d00bb37939b6a8" Dec 04 06:29:01 crc kubenswrapper[4685]: I1204 06:29:01.049353 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9bq7c" Dec 04 06:29:01 crc kubenswrapper[4685]: I1204 06:29:01.177538 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:29:01 crc kubenswrapper[4685]: I1204 06:29:01.183802 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/af46303b-ea28-4ab3-acb7-56a16afa6afd-etc-swift\") pod \"swift-storage-0\" (UID: \"af46303b-ea28-4ab3-acb7-56a16afa6afd\") " pod="openstack/swift-storage-0" Dec 04 06:29:01 crc kubenswrapper[4685]: I1204 06:29:01.421334 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 06:29:02 crc kubenswrapper[4685]: I1204 06:29:02.042004 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 06:29:02 crc kubenswrapper[4685]: I1204 06:29:02.114659 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"dd1f42a68bff31700cc0c13f01d7f7d0efc79ad023c084db0b781501265e353c"} Dec 04 06:29:06 crc kubenswrapper[4685]: I1204 06:29:06.121259 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:29:06 crc kubenswrapper[4685]: I1204 06:29:06.122237 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.173790 4685 generic.go:334] "Generic (PLEG): container finished" podID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerID="d20cedfa26345e0d3ffb056841c394db2b49e8dab26dc2b4f19d7034f3ab834f" exitCode=0 Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.173877 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"777e7995-c510-4d30-bcf1-d479e7cfa432","Type":"ContainerDied","Data":"d20cedfa26345e0d3ffb056841c394db2b49e8dab26dc2b4f19d7034f3ab834f"} Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.176321 4685 generic.go:334] "Generic (PLEG): container finished" podID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerID="a1c9f7e2cb752c2aae9f83d4ccd671a7be6e02092fac5863bf3f397693e89979" exitCode=0 Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.176375 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ff4ac6f-e715-439b-8b4a-2b2715198dab","Type":"ContainerDied","Data":"a1c9f7e2cb752c2aae9f83d4ccd671a7be6e02092fac5863bf3f397693e89979"} Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.773345 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z562k" podUID="2db0269f-8ebe-48f0-81de-081da72da040" containerName="ovn-controller" probeResult="failure" output=< Dec 04 06:29:07 crc kubenswrapper[4685]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 06:29:07 crc kubenswrapper[4685]: > Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.823550 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:29:07 crc kubenswrapper[4685]: I1204 06:29:07.828996 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-wxmhr" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.072358 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z562k-config-v7z56"] Dec 04 06:29:08 crc kubenswrapper[4685]: E1204 06:29:08.072776 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="996bb757-370f-4635-9b6f-923ecd70a8ee" containerName="swift-ring-rebalance" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.072791 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="996bb757-370f-4635-9b6f-923ecd70a8ee" containerName="swift-ring-rebalance" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.072965 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="996bb757-370f-4635-9b6f-923ecd70a8ee" containerName="swift-ring-rebalance" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.073574 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.077602 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.093137 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z562k-config-v7z56"] Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.122579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run-ovn\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.122769 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.122798 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-scripts\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.122820 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-additional-scripts\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.122849 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-log-ovn\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.122871 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsz4s\" (UniqueName: \"kubernetes.io/projected/752a763c-0c5c-4a90-ad6c-056232630807-kube-api-access-vsz4s\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.226402 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.226460 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-scripts\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.226486 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-additional-scripts\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.226509 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-log-ovn\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.226529 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsz4s\" (UniqueName: \"kubernetes.io/projected/752a763c-0c5c-4a90-ad6c-056232630807-kube-api-access-vsz4s\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.226552 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run-ovn\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.227562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run-ovn\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.227638 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.228368 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-additional-scripts\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.228956 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-log-ovn\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.230781 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-scripts\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.258984 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsz4s\" (UniqueName: \"kubernetes.io/projected/752a763c-0c5c-4a90-ad6c-056232630807-kube-api-access-vsz4s\") pod \"ovn-controller-z562k-config-v7z56\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:08 crc kubenswrapper[4685]: I1204 06:29:08.398987 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.251182 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"777e7995-c510-4d30-bcf1-d479e7cfa432","Type":"ContainerStarted","Data":"ff8a6270eb509f97464e48636bef08861603eafd47f756d6a55a49213db69fb7"} Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.254074 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.259483 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"a698afb76ee8ab601544b689d305015fa8de2389735690359d90d28701dd8561"} Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.270643 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ff4ac6f-e715-439b-8b4a-2b2715198dab","Type":"ContainerStarted","Data":"3245c6f54f8a85b6c7f7f8c68af3d97378f0a05f4933a7250df031df4c1593e9"} Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.271121 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.297138 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.215923797 podStartE2EDuration="1m4.297111843s" podCreationTimestamp="2025-12-04 06:28:06 +0000 UTC" firstStartedPulling="2025-12-04 06:28:21.402274947 +0000 UTC m=+1038.590505772" lastFinishedPulling="2025-12-04 06:28:31.483463023 +0000 UTC m=+1048.671693818" observedRunningTime="2025-12-04 06:29:10.286275921 +0000 UTC m=+1087.474506696" watchObservedRunningTime="2025-12-04 06:29:10.297111843 +0000 UTC m=+1087.485342618" Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.316998 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.419673119 podStartE2EDuration="1m4.316977244s" podCreationTimestamp="2025-12-04 06:28:06 +0000 UTC" firstStartedPulling="2025-12-04 06:28:22.867162877 +0000 UTC m=+1040.055393652" lastFinishedPulling="2025-12-04 06:28:31.764466982 +0000 UTC m=+1048.952697777" observedRunningTime="2025-12-04 06:29:10.307735729 +0000 UTC m=+1087.495966504" watchObservedRunningTime="2025-12-04 06:29:10.316977244 +0000 UTC m=+1087.505208019" Dec 04 06:29:10 crc kubenswrapper[4685]: I1204 06:29:10.410102 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z562k-config-v7z56"] Dec 04 06:29:10 crc kubenswrapper[4685]: W1204 06:29:10.426292 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod752a763c_0c5c_4a90_ad6c_056232630807.slice/crio-14768fe96461a9dcaee2e1cbefa84285bbf3e042ebc90a6272a7cc872c4a60f2 WatchSource:0}: Error finding container 14768fe96461a9dcaee2e1cbefa84285bbf3e042ebc90a6272a7cc872c4a60f2: Status 404 returned error can't find the container with id 14768fe96461a9dcaee2e1cbefa84285bbf3e042ebc90a6272a7cc872c4a60f2 Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.287650 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8bkmd" event={"ID":"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0","Type":"ContainerStarted","Data":"0b5ac81183e24c805c82016dcc0273fe2f7185d64132e5578c355a40f2ef4ab0"} Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.291192 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"4496f6aef8271300c67a20a9c3d53c80f29f4438c53d4c2a7f4b9345393e08d5"} Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.291235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"41f551e61e040dbc5e61f0ddb53cd37022c7f911ad16b61e1f01799624c35560"} Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.291246 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"8ab2fb482da345e0d511e4add88e922e0bfdd3696e9ddce5e0d8209e17cb442f"} Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.292847 4685 generic.go:334] "Generic (PLEG): container finished" podID="752a763c-0c5c-4a90-ad6c-056232630807" containerID="1983666a453ada04c22d70ad58c4de6d7f7153e22fd830178d264146af484ab0" exitCode=0 Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.292957 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k-config-v7z56" event={"ID":"752a763c-0c5c-4a90-ad6c-056232630807","Type":"ContainerDied","Data":"1983666a453ada04c22d70ad58c4de6d7f7153e22fd830178d264146af484ab0"} Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.292975 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k-config-v7z56" event={"ID":"752a763c-0c5c-4a90-ad6c-056232630807","Type":"ContainerStarted","Data":"14768fe96461a9dcaee2e1cbefa84285bbf3e042ebc90a6272a7cc872c4a60f2"} Dec 04 06:29:11 crc kubenswrapper[4685]: I1204 06:29:11.322997 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8bkmd" podStartSLOduration=3.3394129 podStartE2EDuration="15.322959241s" podCreationTimestamp="2025-12-04 06:28:56 +0000 UTC" firstStartedPulling="2025-12-04 06:28:58.009912928 +0000 UTC m=+1075.198143713" lastFinishedPulling="2025-12-04 06:29:09.993459279 +0000 UTC m=+1087.181690054" observedRunningTime="2025-12-04 06:29:11.317455358 +0000 UTC m=+1088.505686133" watchObservedRunningTime="2025-12-04 06:29:11.322959241 +0000 UTC m=+1088.511190016" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.306880 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"0572621db88d687696582fa1215ca1cf513646814fbb89b0afcf467480216a41"} Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.307314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"623bcfbad9f69ebae6415ed95085ffc066240fb35296467e9d5944add8a72887"} Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.759290 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.760271 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-z562k" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831579 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-scripts\") pod \"752a763c-0c5c-4a90-ad6c-056232630807\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831650 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run\") pod \"752a763c-0c5c-4a90-ad6c-056232630807\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831735 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-additional-scripts\") pod \"752a763c-0c5c-4a90-ad6c-056232630807\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831764 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run-ovn\") pod \"752a763c-0c5c-4a90-ad6c-056232630807\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831789 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsz4s\" (UniqueName: \"kubernetes.io/projected/752a763c-0c5c-4a90-ad6c-056232630807-kube-api-access-vsz4s\") pod \"752a763c-0c5c-4a90-ad6c-056232630807\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831816 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run" (OuterVolumeSpecName: "var-run") pod "752a763c-0c5c-4a90-ad6c-056232630807" (UID: "752a763c-0c5c-4a90-ad6c-056232630807"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831851 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-log-ovn\") pod \"752a763c-0c5c-4a90-ad6c-056232630807\" (UID: \"752a763c-0c5c-4a90-ad6c-056232630807\") " Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831864 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "752a763c-0c5c-4a90-ad6c-056232630807" (UID: "752a763c-0c5c-4a90-ad6c-056232630807"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.831902 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "752a763c-0c5c-4a90-ad6c-056232630807" (UID: "752a763c-0c5c-4a90-ad6c-056232630807"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.832476 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "752a763c-0c5c-4a90-ad6c-056232630807" (UID: "752a763c-0c5c-4a90-ad6c-056232630807"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.832774 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-scripts" (OuterVolumeSpecName: "scripts") pod "752a763c-0c5c-4a90-ad6c-056232630807" (UID: "752a763c-0c5c-4a90-ad6c-056232630807"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.833467 4685 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.833483 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.833491 4685 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.833499 4685 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/752a763c-0c5c-4a90-ad6c-056232630807-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.833510 4685 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/752a763c-0c5c-4a90-ad6c-056232630807-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.845011 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/752a763c-0c5c-4a90-ad6c-056232630807-kube-api-access-vsz4s" (OuterVolumeSpecName: "kube-api-access-vsz4s") pod "752a763c-0c5c-4a90-ad6c-056232630807" (UID: "752a763c-0c5c-4a90-ad6c-056232630807"). InnerVolumeSpecName "kube-api-access-vsz4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:12 crc kubenswrapper[4685]: I1204 06:29:12.935930 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsz4s\" (UniqueName: \"kubernetes.io/projected/752a763c-0c5c-4a90-ad6c-056232630807-kube-api-access-vsz4s\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.317440 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k-config-v7z56" event={"ID":"752a763c-0c5c-4a90-ad6c-056232630807","Type":"ContainerDied","Data":"14768fe96461a9dcaee2e1cbefa84285bbf3e042ebc90a6272a7cc872c4a60f2"} Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.317493 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14768fe96461a9dcaee2e1cbefa84285bbf3e042ebc90a6272a7cc872c4a60f2" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.317522 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-v7z56" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.323942 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"ef13ea08a24bc0d88959732cacadd43a7942c0d9a02d7523b5453bab05edd17b"} Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.324015 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"4d2ac1a7b1a91c6d66f6308e763d1d03f5e41c9a720960128a139db5401a2e23"} Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.862055 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z562k-config-v7z56"] Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.868514 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z562k-config-v7z56"] Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.989437 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z562k-config-vx944"] Dec 04 06:29:13 crc kubenswrapper[4685]: E1204 06:29:13.989872 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752a763c-0c5c-4a90-ad6c-056232630807" containerName="ovn-config" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.989953 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="752a763c-0c5c-4a90-ad6c-056232630807" containerName="ovn-config" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.990229 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="752a763c-0c5c-4a90-ad6c-056232630807" containerName="ovn-config" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.990993 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:13 crc kubenswrapper[4685]: I1204 06:29:13.992959 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.017007 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z562k-config-vx944"] Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.057654 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-log-ovn\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.057716 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-scripts\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.057739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.057888 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jlc9\" (UniqueName: \"kubernetes.io/projected/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-kube-api-access-4jlc9\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.058066 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run-ovn\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.058121 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-additional-scripts\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.159579 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-additional-scripts\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.159680 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-log-ovn\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.159724 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-scripts\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.159749 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.160019 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-log-ovn\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.160093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jlc9\" (UniqueName: \"kubernetes.io/projected/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-kube-api-access-4jlc9\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.160160 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.160310 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-additional-scripts\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.160479 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run-ovn\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.160825 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run-ovn\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.162093 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-scripts\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.185988 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jlc9\" (UniqueName: \"kubernetes.io/projected/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-kube-api-access-4jlc9\") pod \"ovn-controller-z562k-config-vx944\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.359909 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"640ea0ff8f48506300e3e9e35c27d7610fceb30eb05f976ecc9694295a9db3f1"} Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.360277 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"75499d665ec13afbc322d7c1760bc37cb465b0c4489fd457ea87e41c2f0c22c7"} Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.360291 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"0457a0c187662e4bab6fa75b5311e9fcdb2899a3505b0eb47b314ad806cae53b"} Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.370059 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:14 crc kubenswrapper[4685]: I1204 06:29:14.823530 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z562k-config-vx944"] Dec 04 06:29:14 crc kubenswrapper[4685]: W1204 06:29:14.835291 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a9d9e16_ef51_4ad1_bffe_171ce24e16aa.slice/crio-d9080c63d17e243573d20591533eaf3c641890fbdb24a7d75915118e8b08a113 WatchSource:0}: Error finding container d9080c63d17e243573d20591533eaf3c641890fbdb24a7d75915118e8b08a113: Status 404 returned error can't find the container with id d9080c63d17e243573d20591533eaf3c641890fbdb24a7d75915118e8b08a113 Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.139895 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="752a763c-0c5c-4a90-ad6c-056232630807" path="/var/lib/kubelet/pods/752a763c-0c5c-4a90-ad6c-056232630807/volumes" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.378622 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"379a5065cf87c730099f645474748a70274893238fd3fafddd01bb4cb5a2f45f"} Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.378889 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"51e99786201e8145ebc1dc4f834014e6396c5bd3e36b27219d798d7dd5128bc2"} Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.378904 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"28985bbc5d5745da32bd379a419c5b726b7bfad0510f3f082d1182a834c36aeb"} Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.378913 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"af46303b-ea28-4ab3-acb7-56a16afa6afd","Type":"ContainerStarted","Data":"0f47aec7e03a4f99c0a7ad739502e0620e53918c1e0c5f909e3cd12ec78ef03e"} Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.381752 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k-config-vx944" event={"ID":"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa","Type":"ContainerStarted","Data":"c69fbb5aee85e8c9939d6097e3f18da98a2b6f751df6ea1ef0ec34bf5c237bb8"} Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.381844 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k-config-vx944" event={"ID":"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa","Type":"ContainerStarted","Data":"d9080c63d17e243573d20591533eaf3c641890fbdb24a7d75915118e8b08a113"} Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.424056 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.917306083 podStartE2EDuration="31.424026719s" podCreationTimestamp="2025-12-04 06:28:44 +0000 UTC" firstStartedPulling="2025-12-04 06:29:02.091546467 +0000 UTC m=+1079.279777242" lastFinishedPulling="2025-12-04 06:29:13.598267103 +0000 UTC m=+1090.786497878" observedRunningTime="2025-12-04 06:29:15.417168915 +0000 UTC m=+1092.605399690" watchObservedRunningTime="2025-12-04 06:29:15.424026719 +0000 UTC m=+1092.612257514" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.452966 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z562k-config-vx944" podStartSLOduration=2.452946149 podStartE2EDuration="2.452946149s" podCreationTimestamp="2025-12-04 06:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:15.446529088 +0000 UTC m=+1092.634759893" watchObservedRunningTime="2025-12-04 06:29:15.452946149 +0000 UTC m=+1092.641176924" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.718942 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-qp7fx"] Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.720700 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.728150 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.733298 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-qp7fx"] Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.892793 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.892911 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-config\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.892959 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.893004 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.893036 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.893062 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgm6s\" (UniqueName: \"kubernetes.io/projected/8fa46156-0695-4259-a8e6-8325a48ccb08-kube-api-access-wgm6s\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.994126 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.994192 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.994218 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.994240 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgm6s\" (UniqueName: \"kubernetes.io/projected/8fa46156-0695-4259-a8e6-8325a48ccb08-kube-api-access-wgm6s\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.994323 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.994385 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-config\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.995239 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.995239 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.995306 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.995462 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-config\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:15 crc kubenswrapper[4685]: I1204 06:29:15.995534 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:16 crc kubenswrapper[4685]: I1204 06:29:16.025447 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgm6s\" (UniqueName: \"kubernetes.io/projected/8fa46156-0695-4259-a8e6-8325a48ccb08-kube-api-access-wgm6s\") pod \"dnsmasq-dns-5c79d794d7-qp7fx\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:16 crc kubenswrapper[4685]: I1204 06:29:16.041081 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:16 crc kubenswrapper[4685]: I1204 06:29:16.400067 4685 generic.go:334] "Generic (PLEG): container finished" podID="5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" containerID="c69fbb5aee85e8c9939d6097e3f18da98a2b6f751df6ea1ef0ec34bf5c237bb8" exitCode=0 Dec 04 06:29:16 crc kubenswrapper[4685]: I1204 06:29:16.402315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z562k-config-vx944" event={"ID":"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa","Type":"ContainerDied","Data":"c69fbb5aee85e8c9939d6097e3f18da98a2b6f751df6ea1ef0ec34bf5c237bb8"} Dec 04 06:29:16 crc kubenswrapper[4685]: I1204 06:29:16.543196 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-qp7fx"] Dec 04 06:29:16 crc kubenswrapper[4685]: W1204 06:29:16.552393 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa46156_0695_4259_a8e6_8325a48ccb08.slice/crio-a02cea9ebda505d5c162b9df18fdd3a756b5ad2b78b094d5504d5588395a2b12 WatchSource:0}: Error finding container a02cea9ebda505d5c162b9df18fdd3a756b5ad2b78b094d5504d5588395a2b12: Status 404 returned error can't find the container with id a02cea9ebda505d5c162b9df18fdd3a756b5ad2b78b094d5504d5588395a2b12 Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.410138 4685 generic.go:334] "Generic (PLEG): container finished" podID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerID="777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76" exitCode=0 Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.410243 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" event={"ID":"8fa46156-0695-4259-a8e6-8325a48ccb08","Type":"ContainerDied","Data":"777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76"} Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.410554 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" event={"ID":"8fa46156-0695-4259-a8e6-8325a48ccb08","Type":"ContainerStarted","Data":"a02cea9ebda505d5c162b9df18fdd3a756b5ad2b78b094d5504d5588395a2b12"} Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.412708 4685 generic.go:334] "Generic (PLEG): container finished" podID="2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" containerID="0b5ac81183e24c805c82016dcc0273fe2f7185d64132e5578c355a40f2ef4ab0" exitCode=0 Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.412848 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8bkmd" event={"ID":"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0","Type":"ContainerDied","Data":"0b5ac81183e24c805c82016dcc0273fe2f7185d64132e5578c355a40f2ef4ab0"} Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.725481 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.832979 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-scripts\") pod \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833058 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run-ovn\") pod \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833098 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run\") pod \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833192 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jlc9\" (UniqueName: \"kubernetes.io/projected/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-kube-api-access-4jlc9\") pod \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833369 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-additional-scripts\") pod \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833397 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-log-ovn\") pod \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\" (UID: \"5a9d9e16-ef51-4ad1-bffe-171ce24e16aa\") " Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833572 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run" (OuterVolumeSpecName: "var-run") pod "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" (UID: "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.833653 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" (UID: "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.834088 4685 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.834106 4685 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.834135 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" (UID: "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.834428 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" (UID: "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.834642 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-scripts" (OuterVolumeSpecName: "scripts") pod "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" (UID: "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.838610 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-kube-api-access-4jlc9" (OuterVolumeSpecName: "kube-api-access-4jlc9") pod "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" (UID: "5a9d9e16-ef51-4ad1-bffe-171ce24e16aa"). InnerVolumeSpecName "kube-api-access-4jlc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.894997 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z562k-config-vx944"] Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.900648 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z562k-config-vx944"] Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.935426 4685 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.935827 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.935936 4685 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:17 crc kubenswrapper[4685]: I1204 06:29:17.936016 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jlc9\" (UniqueName: \"kubernetes.io/projected/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa-kube-api-access-4jlc9\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:18 crc kubenswrapper[4685]: I1204 06:29:18.423343 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" event={"ID":"8fa46156-0695-4259-a8e6-8325a48ccb08","Type":"ContainerStarted","Data":"c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55"} Dec 04 06:29:18 crc kubenswrapper[4685]: I1204 06:29:18.423753 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:18 crc kubenswrapper[4685]: I1204 06:29:18.426231 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9080c63d17e243573d20591533eaf3c641890fbdb24a7d75915118e8b08a113" Dec 04 06:29:18 crc kubenswrapper[4685]: I1204 06:29:18.426252 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z562k-config-vx944" Dec 04 06:29:18 crc kubenswrapper[4685]: I1204 06:29:18.911885 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8bkmd" Dec 04 06:29:18 crc kubenswrapper[4685]: I1204 06:29:18.928971 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" podStartSLOduration=3.92895285 podStartE2EDuration="3.92895285s" podCreationTimestamp="2025-12-04 06:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:18.467185343 +0000 UTC m=+1095.655416128" watchObservedRunningTime="2025-12-04 06:29:18.92895285 +0000 UTC m=+1096.117183625" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.056907 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-combined-ca-bundle\") pod \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.057047 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-db-sync-config-data\") pod \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.057088 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz2jc\" (UniqueName: \"kubernetes.io/projected/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-kube-api-access-wz2jc\") pod \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.057191 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-config-data\") pod \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\" (UID: \"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0\") " Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.063043 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" (UID: "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.065369 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-kube-api-access-wz2jc" (OuterVolumeSpecName: "kube-api-access-wz2jc") pod "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" (UID: "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0"). InnerVolumeSpecName "kube-api-access-wz2jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.107380 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" (UID: "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.133329 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-config-data" (OuterVolumeSpecName: "config-data") pod "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" (UID: "2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.162044 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.162087 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.162102 4685 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.162114 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz2jc\" (UniqueName: \"kubernetes.io/projected/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0-kube-api-access-wz2jc\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.174603 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" path="/var/lib/kubelet/pods/5a9d9e16-ef51-4ad1-bffe-171ce24e16aa/volumes" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.445638 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8bkmd" Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.446601 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8bkmd" event={"ID":"2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0","Type":"ContainerDied","Data":"f68aab5abc5f4913f6afc3186cff505af58115c49e859e5a182b24b31ac7eb6c"} Dec 04 06:29:19 crc kubenswrapper[4685]: I1204 06:29:19.446637 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f68aab5abc5f4913f6afc3186cff505af58115c49e859e5a182b24b31ac7eb6c" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.071546 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-qp7fx"] Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.122937 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9wqf7"] Dec 04 06:29:20 crc kubenswrapper[4685]: E1204 06:29:20.123320 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" containerName="ovn-config" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.123341 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" containerName="ovn-config" Dec 04 06:29:20 crc kubenswrapper[4685]: E1204 06:29:20.123370 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" containerName="glance-db-sync" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.123379 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" containerName="glance-db-sync" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.123668 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9d9e16-ef51-4ad1-bffe-171ce24e16aa" containerName="ovn-config" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.123697 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" containerName="glance-db-sync" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.124591 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.147447 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9wqf7"] Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.190956 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlwll\" (UniqueName: \"kubernetes.io/projected/991d5443-6323-48ba-9a55-832b4d80820b-kube-api-access-nlwll\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.191013 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.191051 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.191098 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-config\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.191129 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.191152 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.292639 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.292721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.292764 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-config\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.292793 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.292814 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.294287 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.294322 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-config\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.294348 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.294499 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlwll\" (UniqueName: \"kubernetes.io/projected/991d5443-6323-48ba-9a55-832b4d80820b-kube-api-access-nlwll\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.294804 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.295879 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.311562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlwll\" (UniqueName: \"kubernetes.io/projected/991d5443-6323-48ba-9a55-832b4d80820b-kube-api-access-nlwll\") pod \"dnsmasq-dns-5f59b8f679-9wqf7\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.441087 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.450643 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerName="dnsmasq-dns" containerID="cri-o://c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55" gracePeriod=10 Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.754384 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9wqf7"] Dec 04 06:29:20 crc kubenswrapper[4685]: I1204 06:29:20.914104 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.008006 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgm6s\" (UniqueName: \"kubernetes.io/projected/8fa46156-0695-4259-a8e6-8325a48ccb08-kube-api-access-wgm6s\") pod \"8fa46156-0695-4259-a8e6-8325a48ccb08\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.008172 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-sb\") pod \"8fa46156-0695-4259-a8e6-8325a48ccb08\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.008357 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-config\") pod \"8fa46156-0695-4259-a8e6-8325a48ccb08\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.008543 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-nb\") pod \"8fa46156-0695-4259-a8e6-8325a48ccb08\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.008684 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-svc\") pod \"8fa46156-0695-4259-a8e6-8325a48ccb08\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.008802 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-swift-storage-0\") pod \"8fa46156-0695-4259-a8e6-8325a48ccb08\" (UID: \"8fa46156-0695-4259-a8e6-8325a48ccb08\") " Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.015608 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa46156-0695-4259-a8e6-8325a48ccb08-kube-api-access-wgm6s" (OuterVolumeSpecName: "kube-api-access-wgm6s") pod "8fa46156-0695-4259-a8e6-8325a48ccb08" (UID: "8fa46156-0695-4259-a8e6-8325a48ccb08"). InnerVolumeSpecName "kube-api-access-wgm6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.061561 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-config" (OuterVolumeSpecName: "config") pod "8fa46156-0695-4259-a8e6-8325a48ccb08" (UID: "8fa46156-0695-4259-a8e6-8325a48ccb08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.066255 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8fa46156-0695-4259-a8e6-8325a48ccb08" (UID: "8fa46156-0695-4259-a8e6-8325a48ccb08"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.069969 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8fa46156-0695-4259-a8e6-8325a48ccb08" (UID: "8fa46156-0695-4259-a8e6-8325a48ccb08"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.074335 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8fa46156-0695-4259-a8e6-8325a48ccb08" (UID: "8fa46156-0695-4259-a8e6-8325a48ccb08"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.078604 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8fa46156-0695-4259-a8e6-8325a48ccb08" (UID: "8fa46156-0695-4259-a8e6-8325a48ccb08"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.110701 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgm6s\" (UniqueName: \"kubernetes.io/projected/8fa46156-0695-4259-a8e6-8325a48ccb08-kube-api-access-wgm6s\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.110930 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.110943 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.110954 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.110967 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.110977 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa46156-0695-4259-a8e6-8325a48ccb08-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.465295 4685 generic.go:334] "Generic (PLEG): container finished" podID="991d5443-6323-48ba-9a55-832b4d80820b" containerID="b7effbed9be6c6720b522757019ea495f31b64e5e9664c9384e58554e62eff9c" exitCode=0 Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.465379 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" event={"ID":"991d5443-6323-48ba-9a55-832b4d80820b","Type":"ContainerDied","Data":"b7effbed9be6c6720b522757019ea495f31b64e5e9664c9384e58554e62eff9c"} Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.466456 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" event={"ID":"991d5443-6323-48ba-9a55-832b4d80820b","Type":"ContainerStarted","Data":"88b62b262eca01b36fcbd31c9e70952a2ce38874d6a7f906bc1109e51f732f10"} Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.469872 4685 generic.go:334] "Generic (PLEG): container finished" podID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerID="c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55" exitCode=0 Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.469938 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" event={"ID":"8fa46156-0695-4259-a8e6-8325a48ccb08","Type":"ContainerDied","Data":"c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55"} Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.469979 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" event={"ID":"8fa46156-0695-4259-a8e6-8325a48ccb08","Type":"ContainerDied","Data":"a02cea9ebda505d5c162b9df18fdd3a756b5ad2b78b094d5504d5588395a2b12"} Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.470057 4685 scope.go:117] "RemoveContainer" containerID="c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.470317 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-qp7fx" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.535726 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-qp7fx"] Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.551488 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-qp7fx"] Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.639827 4685 scope.go:117] "RemoveContainer" containerID="777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.668819 4685 scope.go:117] "RemoveContainer" containerID="c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55" Dec 04 06:29:21 crc kubenswrapper[4685]: E1204 06:29:21.669520 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55\": container with ID starting with c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55 not found: ID does not exist" containerID="c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.669575 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55"} err="failed to get container status \"c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55\": rpc error: code = NotFound desc = could not find container \"c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55\": container with ID starting with c8d9e035a0b3b91e87a8ba42db7376593b3a47ca171a2c659407f8c046541f55 not found: ID does not exist" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.669613 4685 scope.go:117] "RemoveContainer" containerID="777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76" Dec 04 06:29:21 crc kubenswrapper[4685]: E1204 06:29:21.670113 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76\": container with ID starting with 777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76 not found: ID does not exist" containerID="777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76" Dec 04 06:29:21 crc kubenswrapper[4685]: I1204 06:29:21.670172 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76"} err="failed to get container status \"777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76\": rpc error: code = NotFound desc = could not find container \"777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76\": container with ID starting with 777dbfdcb692c97f4f3f538655e67f9e68b0e329fa34e753859596e078419a76 not found: ID does not exist" Dec 04 06:29:22 crc kubenswrapper[4685]: I1204 06:29:22.483236 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" event={"ID":"991d5443-6323-48ba-9a55-832b4d80820b","Type":"ContainerStarted","Data":"179ee4c4f3bb1d39611f2e3d12131ac9da868bf495deca3146a5b9cc40bd0ed6"} Dec 04 06:29:22 crc kubenswrapper[4685]: I1204 06:29:22.483710 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:22 crc kubenswrapper[4685]: I1204 06:29:22.516086 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" podStartSLOduration=2.516065368 podStartE2EDuration="2.516065368s" podCreationTimestamp="2025-12-04 06:29:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:22.508902135 +0000 UTC m=+1099.697132920" watchObservedRunningTime="2025-12-04 06:29:22.516065368 +0000 UTC m=+1099.704296143" Dec 04 06:29:23 crc kubenswrapper[4685]: I1204 06:29:23.143214 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" path="/var/lib/kubelet/pods/8fa46156-0695-4259-a8e6-8325a48ccb08/volumes" Dec 04 06:29:27 crc kubenswrapper[4685]: I1204 06:29:27.878708 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.217164 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.381462 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tmptx"] Dec 04 06:29:28 crc kubenswrapper[4685]: E1204 06:29:28.381802 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerName="init" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.381817 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerName="init" Dec 04 06:29:28 crc kubenswrapper[4685]: E1204 06:29:28.381834 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerName="dnsmasq-dns" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.381841 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerName="dnsmasq-dns" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.382015 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa46156-0695-4259-a8e6-8325a48ccb08" containerName="dnsmasq-dns" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.382576 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.392477 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tmptx"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.395424 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989f07ae-40ac-4b8d-91de-167df369b845-operator-scripts\") pod \"cinder-db-create-tmptx\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.395503 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6j7q\" (UniqueName: \"kubernetes.io/projected/989f07ae-40ac-4b8d-91de-167df369b845-kube-api-access-p6j7q\") pod \"cinder-db-create-tmptx\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.425157 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-41f6-account-create-update-ctjs8"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.426471 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.428993 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.484644 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-v6sh8"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.485995 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.492533 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-41f6-account-create-update-ctjs8"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.496971 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rqrt\" (UniqueName: \"kubernetes.io/projected/3a0c2c44-843d-4750-abbe-b4a9effe469f-kube-api-access-5rqrt\") pod \"cinder-41f6-account-create-update-ctjs8\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.497013 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6j7q\" (UniqueName: \"kubernetes.io/projected/989f07ae-40ac-4b8d-91de-167df369b845-kube-api-access-p6j7q\") pod \"cinder-db-create-tmptx\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.497051 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b071b469-ce94-4165-a933-55f9f8f36abe-operator-scripts\") pod \"barbican-db-create-v6sh8\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.497099 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a0c2c44-843d-4750-abbe-b4a9effe469f-operator-scripts\") pod \"cinder-41f6-account-create-update-ctjs8\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.497137 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbkwp\" (UniqueName: \"kubernetes.io/projected/b071b469-ce94-4165-a933-55f9f8f36abe-kube-api-access-vbkwp\") pod \"barbican-db-create-v6sh8\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.497172 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989f07ae-40ac-4b8d-91de-167df369b845-operator-scripts\") pod \"cinder-db-create-tmptx\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.497857 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989f07ae-40ac-4b8d-91de-167df369b845-operator-scripts\") pod \"cinder-db-create-tmptx\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.511704 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v6sh8"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.527264 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6j7q\" (UniqueName: \"kubernetes.io/projected/989f07ae-40ac-4b8d-91de-167df369b845-kube-api-access-p6j7q\") pod \"cinder-db-create-tmptx\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.597793 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b071b469-ce94-4165-a933-55f9f8f36abe-operator-scripts\") pod \"barbican-db-create-v6sh8\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.597862 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a0c2c44-843d-4750-abbe-b4a9effe469f-operator-scripts\") pod \"cinder-41f6-account-create-update-ctjs8\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.597897 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbkwp\" (UniqueName: \"kubernetes.io/projected/b071b469-ce94-4165-a933-55f9f8f36abe-kube-api-access-vbkwp\") pod \"barbican-db-create-v6sh8\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.597946 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rqrt\" (UniqueName: \"kubernetes.io/projected/3a0c2c44-843d-4750-abbe-b4a9effe469f-kube-api-access-5rqrt\") pod \"cinder-41f6-account-create-update-ctjs8\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.598906 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b071b469-ce94-4165-a933-55f9f8f36abe-operator-scripts\") pod \"barbican-db-create-v6sh8\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.599399 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a0c2c44-843d-4750-abbe-b4a9effe469f-operator-scripts\") pod \"cinder-41f6-account-create-update-ctjs8\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.614671 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbkwp\" (UniqueName: \"kubernetes.io/projected/b071b469-ce94-4165-a933-55f9f8f36abe-kube-api-access-vbkwp\") pod \"barbican-db-create-v6sh8\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.615458 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rqrt\" (UniqueName: \"kubernetes.io/projected/3a0c2c44-843d-4750-abbe-b4a9effe469f-kube-api-access-5rqrt\") pod \"cinder-41f6-account-create-update-ctjs8\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.686637 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-f5f0-account-create-update-gl7hp"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.687649 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: W1204 06:29:28.690108 4685 reflector.go:561] object-"openstack"/"barbican-db-secret": failed to list *v1.Secret: secrets "barbican-db-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 04 06:29:28 crc kubenswrapper[4685]: E1204 06:29:28.690144 4685 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"barbican-db-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"barbican-db-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.698907 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.699765 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f5f0-account-create-update-gl7hp"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.699796 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwmfg\" (UniqueName: \"kubernetes.io/projected/ea062396-352c-4a83-8745-5bec1a9284bc-kube-api-access-dwmfg\") pod \"barbican-f5f0-account-create-update-gl7hp\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.699957 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea062396-352c-4a83-8745-5bec1a9284bc-operator-scripts\") pod \"barbican-f5f0-account-create-update-gl7hp\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.746382 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.776522 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-sh2ng"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.777698 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.798479 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-sh2ng"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.803044 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwmfg\" (UniqueName: \"kubernetes.io/projected/ea062396-352c-4a83-8745-5bec1a9284bc-kube-api-access-dwmfg\") pod \"barbican-f5f0-account-create-update-gl7hp\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.803188 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea062396-352c-4a83-8745-5bec1a9284bc-operator-scripts\") pod \"barbican-f5f0-account-create-update-gl7hp\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.803921 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea062396-352c-4a83-8745-5bec1a9284bc-operator-scripts\") pod \"barbican-f5f0-account-create-update-gl7hp\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.805626 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.830600 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwmfg\" (UniqueName: \"kubernetes.io/projected/ea062396-352c-4a83-8745-5bec1a9284bc-kube-api-access-dwmfg\") pod \"barbican-f5f0-account-create-update-gl7hp\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.867985 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wlwr5"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.885253 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.887710 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wlwr5"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.889008 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.889188 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.889294 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c6l2p" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.889455 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.905459 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-operator-scripts\") pod \"neutron-db-create-sh2ng\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.905509 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb7q7\" (UniqueName: \"kubernetes.io/projected/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-kube-api-access-tb7q7\") pod \"neutron-db-create-sh2ng\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.918514 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5b79-account-create-update-bdgk2"] Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.919728 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.925732 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 06:29:28 crc kubenswrapper[4685]: I1204 06:29:28.943562 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5b79-account-create-update-bdgk2"] Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.008550 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-config-data\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.008630 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-operator-scripts\") pod \"neutron-db-create-sh2ng\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.008670 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2sp8\" (UniqueName: \"kubernetes.io/projected/bd467839-0575-42ac-98fd-4e6d2cb3ceba-kube-api-access-v2sp8\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.008692 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb7q7\" (UniqueName: \"kubernetes.io/projected/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-kube-api-access-tb7q7\") pod \"neutron-db-create-sh2ng\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.008721 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-combined-ca-bundle\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.010019 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-operator-scripts\") pod \"neutron-db-create-sh2ng\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.014849 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.029179 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb7q7\" (UniqueName: \"kubernetes.io/projected/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-kube-api-access-tb7q7\") pod \"neutron-db-create-sh2ng\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.110421 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8tkg\" (UniqueName: \"kubernetes.io/projected/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-kube-api-access-q8tkg\") pod \"neutron-5b79-account-create-update-bdgk2\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.110537 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-config-data\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.110586 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2sp8\" (UniqueName: \"kubernetes.io/projected/bd467839-0575-42ac-98fd-4e6d2cb3ceba-kube-api-access-v2sp8\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.110612 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-combined-ca-bundle\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.110644 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-operator-scripts\") pod \"neutron-5b79-account-create-update-bdgk2\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.117823 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-config-data\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.118929 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-combined-ca-bundle\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.129938 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2sp8\" (UniqueName: \"kubernetes.io/projected/bd467839-0575-42ac-98fd-4e6d2cb3ceba-kube-api-access-v2sp8\") pod \"keystone-db-sync-wlwr5\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.136792 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.206806 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.212067 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-operator-scripts\") pod \"neutron-5b79-account-create-update-bdgk2\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.212129 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8tkg\" (UniqueName: \"kubernetes.io/projected/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-kube-api-access-q8tkg\") pod \"neutron-5b79-account-create-update-bdgk2\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.213236 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-operator-scripts\") pod \"neutron-5b79-account-create-update-bdgk2\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.235400 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8tkg\" (UniqueName: \"kubernetes.io/projected/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-kube-api-access-q8tkg\") pod \"neutron-5b79-account-create-update-bdgk2\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.263913 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.299204 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tmptx"] Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.375699 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v6sh8"] Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.449897 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-41f6-account-create-update-ctjs8"] Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.611243 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tmptx" event={"ID":"989f07ae-40ac-4b8d-91de-167df369b845","Type":"ContainerStarted","Data":"c8acecd5105f329c79e8bba24f0d7621425900766ea61455668a6c506068bd92"} Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.644694 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f5f0-account-create-update-gl7hp"] Dec 04 06:29:29 crc kubenswrapper[4685]: E1204 06:29:29.779651 4685 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.230:39102->38.102.83.230:44137: write tcp 38.102.83.230:39102->38.102.83.230:44137: write: broken pipe Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.866366 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-sh2ng"] Dec 04 06:29:29 crc kubenswrapper[4685]: I1204 06:29:29.958825 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.012680 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wlwr5"] Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.131345 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5b79-account-create-update-bdgk2"] Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.442607 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.501329 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-6bsx8"] Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.505773 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerName="dnsmasq-dns" containerID="cri-o://9dd67008de7ea673e410a7b4e0869ffe00a03de06e4557b6f96c2f152d815e7c" gracePeriod=10 Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.623224 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v6sh8" event={"ID":"b071b469-ce94-4165-a933-55f9f8f36abe","Type":"ContainerStarted","Data":"a1e667d6c16d50648de895728817216cb37996d3d472fb66b2603e40268dffed"} Dec 04 06:29:30 crc kubenswrapper[4685]: I1204 06:29:30.628278 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-41f6-account-create-update-ctjs8" event={"ID":"3a0c2c44-843d-4750-abbe-b4a9effe469f","Type":"ContainerStarted","Data":"2b1fd4519b10be5b5808f75f16d2a4f54a0dc3cdec658767958bb4e135b57166"} Dec 04 06:29:31 crc kubenswrapper[4685]: I1204 06:29:31.641079 4685 generic.go:334] "Generic (PLEG): container finished" podID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerID="9dd67008de7ea673e410a7b4e0869ffe00a03de06e4557b6f96c2f152d815e7c" exitCode=0 Dec 04 06:29:31 crc kubenswrapper[4685]: I1204 06:29:31.641136 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" event={"ID":"de4c5c56-929c-47b1-a488-f9eea7edeccc","Type":"ContainerDied","Data":"9dd67008de7ea673e410a7b4e0869ffe00a03de06e4557b6f96c2f152d815e7c"} Dec 04 06:29:32 crc kubenswrapper[4685]: I1204 06:29:32.653200 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5b79-account-create-update-bdgk2" event={"ID":"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e","Type":"ContainerStarted","Data":"b1a2dbfb0c8e0315c57d84bef8424c27bb21c5fd60752e5f3714c967bede54af"} Dec 04 06:29:32 crc kubenswrapper[4685]: I1204 06:29:32.654286 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sh2ng" event={"ID":"bfbaf5e7-e351-4c00-83e1-a3131589e9d9","Type":"ContainerStarted","Data":"36ef7347921ee69c311aae113c461b89b0087623dd419c8938e4033ee0317836"} Dec 04 06:29:32 crc kubenswrapper[4685]: I1204 06:29:32.656022 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f5f0-account-create-update-gl7hp" event={"ID":"ea062396-352c-4a83-8745-5bec1a9284bc","Type":"ContainerStarted","Data":"d1a395127a3487a29d8040e62b2f5b4fbcdaacd28c67c213fe9f49f4ea04ebd0"} Dec 04 06:29:32 crc kubenswrapper[4685]: I1204 06:29:32.657822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tmptx" event={"ID":"989f07ae-40ac-4b8d-91de-167df369b845","Type":"ContainerStarted","Data":"93bd3f5f888caee883654231951f728342202843c20f1d659213346365251944"} Dec 04 06:29:32 crc kubenswrapper[4685]: I1204 06:29:32.658888 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wlwr5" event={"ID":"bd467839-0575-42ac-98fd-4e6d2cb3ceba","Type":"ContainerStarted","Data":"35d8ccf908e59c3de1846f3af8c8d84efeaba4612adf09b224bbd30a0ee6b80d"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.670646 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" event={"ID":"de4c5c56-929c-47b1-a488-f9eea7edeccc","Type":"ContainerDied","Data":"8993fe3de7909e5d68edd6aa502520fae46a0622c78d34b6af3405418aee427a"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.670973 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8993fe3de7909e5d68edd6aa502520fae46a0622c78d34b6af3405418aee427a" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.674283 4685 generic.go:334] "Generic (PLEG): container finished" podID="ea062396-352c-4a83-8745-5bec1a9284bc" containerID="6fd7328ae0af3c59a457ba4a7e667dfc305438d00100afc19cf6fa4b2bc15cb0" exitCode=0 Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.674349 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f5f0-account-create-update-gl7hp" event={"ID":"ea062396-352c-4a83-8745-5bec1a9284bc","Type":"ContainerDied","Data":"6fd7328ae0af3c59a457ba4a7e667dfc305438d00100afc19cf6fa4b2bc15cb0"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.677901 4685 generic.go:334] "Generic (PLEG): container finished" podID="3a0c2c44-843d-4750-abbe-b4a9effe469f" containerID="4dd65d877c9d947c1e27c08f400d71a542c2120ace1d3794a568be83e96f8e79" exitCode=0 Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.678068 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-41f6-account-create-update-ctjs8" event={"ID":"3a0c2c44-843d-4750-abbe-b4a9effe469f","Type":"ContainerDied","Data":"4dd65d877c9d947c1e27c08f400d71a542c2120ace1d3794a568be83e96f8e79"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.681651 4685 generic.go:334] "Generic (PLEG): container finished" podID="989f07ae-40ac-4b8d-91de-167df369b845" containerID="93bd3f5f888caee883654231951f728342202843c20f1d659213346365251944" exitCode=0 Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.681713 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tmptx" event={"ID":"989f07ae-40ac-4b8d-91de-167df369b845","Type":"ContainerDied","Data":"93bd3f5f888caee883654231951f728342202843c20f1d659213346365251944"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.684340 4685 generic.go:334] "Generic (PLEG): container finished" podID="b071b469-ce94-4165-a933-55f9f8f36abe" containerID="608a3a642b882aef1cb59d4b14f807074834a50894a292a70d5a1860a04f9b6a" exitCode=0 Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.684380 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v6sh8" event={"ID":"b071b469-ce94-4165-a933-55f9f8f36abe","Type":"ContainerDied","Data":"608a3a642b882aef1cb59d4b14f807074834a50894a292a70d5a1860a04f9b6a"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.687014 4685 generic.go:334] "Generic (PLEG): container finished" podID="dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" containerID="0489c7258dea7c5d4594598701151954700bf5703a38d3a569a79c1ec24b8643" exitCode=0 Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.687067 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5b79-account-create-update-bdgk2" event={"ID":"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e","Type":"ContainerDied","Data":"0489c7258dea7c5d4594598701151954700bf5703a38d3a569a79c1ec24b8643"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.692874 4685 generic.go:334] "Generic (PLEG): container finished" podID="bfbaf5e7-e351-4c00-83e1-a3131589e9d9" containerID="7343eab1c44d1c3d6cfb8caa591452bbf50e8639bce0115f5b2bb48413586cf9" exitCode=0 Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.692942 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sh2ng" event={"ID":"bfbaf5e7-e351-4c00-83e1-a3131589e9d9","Type":"ContainerDied","Data":"7343eab1c44d1c3d6cfb8caa591452bbf50e8639bce0115f5b2bb48413586cf9"} Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.762024 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.834522 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-sb\") pod \"de4c5c56-929c-47b1-a488-f9eea7edeccc\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.834624 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-config\") pod \"de4c5c56-929c-47b1-a488-f9eea7edeccc\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.834757 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-dns-svc\") pod \"de4c5c56-929c-47b1-a488-f9eea7edeccc\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.834845 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-nb\") pod \"de4c5c56-929c-47b1-a488-f9eea7edeccc\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.834906 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kfxh\" (UniqueName: \"kubernetes.io/projected/de4c5c56-929c-47b1-a488-f9eea7edeccc-kube-api-access-9kfxh\") pod \"de4c5c56-929c-47b1-a488-f9eea7edeccc\" (UID: \"de4c5c56-929c-47b1-a488-f9eea7edeccc\") " Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.843858 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4c5c56-929c-47b1-a488-f9eea7edeccc-kube-api-access-9kfxh" (OuterVolumeSpecName: "kube-api-access-9kfxh") pod "de4c5c56-929c-47b1-a488-f9eea7edeccc" (UID: "de4c5c56-929c-47b1-a488-f9eea7edeccc"). InnerVolumeSpecName "kube-api-access-9kfxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.879661 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de4c5c56-929c-47b1-a488-f9eea7edeccc" (UID: "de4c5c56-929c-47b1-a488-f9eea7edeccc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.880115 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de4c5c56-929c-47b1-a488-f9eea7edeccc" (UID: "de4c5c56-929c-47b1-a488-f9eea7edeccc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.893308 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de4c5c56-929c-47b1-a488-f9eea7edeccc" (UID: "de4c5c56-929c-47b1-a488-f9eea7edeccc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.902434 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-config" (OuterVolumeSpecName: "config") pod "de4c5c56-929c-47b1-a488-f9eea7edeccc" (UID: "de4c5c56-929c-47b1-a488-f9eea7edeccc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.937335 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.937371 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kfxh\" (UniqueName: \"kubernetes.io/projected/de4c5c56-929c-47b1-a488-f9eea7edeccc-kube-api-access-9kfxh\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.937382 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.937393 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:33 crc kubenswrapper[4685]: I1204 06:29:33.937403 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de4c5c56-929c-47b1-a488-f9eea7edeccc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:34 crc kubenswrapper[4685]: I1204 06:29:34.701865 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-6bsx8" Dec 04 06:29:34 crc kubenswrapper[4685]: I1204 06:29:34.748011 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-6bsx8"] Dec 04 06:29:34 crc kubenswrapper[4685]: I1204 06:29:34.748072 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-6bsx8"] Dec 04 06:29:35 crc kubenswrapper[4685]: I1204 06:29:35.145953 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" path="/var/lib/kubelet/pods/de4c5c56-929c-47b1-a488-f9eea7edeccc/volumes" Dec 04 06:29:36 crc kubenswrapper[4685]: I1204 06:29:36.121705 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:29:36 crc kubenswrapper[4685]: I1204 06:29:36.121772 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.313074 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.318514 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.323729 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.395636 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.404502 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410663 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbkwp\" (UniqueName: \"kubernetes.io/projected/b071b469-ce94-4165-a933-55f9f8f36abe-kube-api-access-vbkwp\") pod \"b071b469-ce94-4165-a933-55f9f8f36abe\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410723 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rqrt\" (UniqueName: \"kubernetes.io/projected/3a0c2c44-843d-4750-abbe-b4a9effe469f-kube-api-access-5rqrt\") pod \"3a0c2c44-843d-4750-abbe-b4a9effe469f\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410788 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8tkg\" (UniqueName: \"kubernetes.io/projected/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-kube-api-access-q8tkg\") pod \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410820 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-operator-scripts\") pod \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\" (UID: \"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410852 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwmfg\" (UniqueName: \"kubernetes.io/projected/ea062396-352c-4a83-8745-5bec1a9284bc-kube-api-access-dwmfg\") pod \"ea062396-352c-4a83-8745-5bec1a9284bc\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410866 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a0c2c44-843d-4750-abbe-b4a9effe469f-operator-scripts\") pod \"3a0c2c44-843d-4750-abbe-b4a9effe469f\" (UID: \"3a0c2c44-843d-4750-abbe-b4a9effe469f\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410888 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b071b469-ce94-4165-a933-55f9f8f36abe-operator-scripts\") pod \"b071b469-ce94-4165-a933-55f9f8f36abe\" (UID: \"b071b469-ce94-4165-a933-55f9f8f36abe\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.410944 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea062396-352c-4a83-8745-5bec1a9284bc-operator-scripts\") pod \"ea062396-352c-4a83-8745-5bec1a9284bc\" (UID: \"ea062396-352c-4a83-8745-5bec1a9284bc\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.412128 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea062396-352c-4a83-8745-5bec1a9284bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea062396-352c-4a83-8745-5bec1a9284bc" (UID: "ea062396-352c-4a83-8745-5bec1a9284bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.412329 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a0c2c44-843d-4750-abbe-b4a9effe469f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a0c2c44-843d-4750-abbe-b4a9effe469f" (UID: "3a0c2c44-843d-4750-abbe-b4a9effe469f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.412585 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" (UID: "dabcb5c7-61d1-46eb-9c15-c14a1e0df79e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.412616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b071b469-ce94-4165-a933-55f9f8f36abe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b071b469-ce94-4165-a933-55f9f8f36abe" (UID: "b071b469-ce94-4165-a933-55f9f8f36abe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.415894 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b071b469-ce94-4165-a933-55f9f8f36abe-kube-api-access-vbkwp" (OuterVolumeSpecName: "kube-api-access-vbkwp") pod "b071b469-ce94-4165-a933-55f9f8f36abe" (UID: "b071b469-ce94-4165-a933-55f9f8f36abe"). InnerVolumeSpecName "kube-api-access-vbkwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.416369 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-kube-api-access-q8tkg" (OuterVolumeSpecName: "kube-api-access-q8tkg") pod "dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" (UID: "dabcb5c7-61d1-46eb-9c15-c14a1e0df79e"). InnerVolumeSpecName "kube-api-access-q8tkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.430590 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea062396-352c-4a83-8745-5bec1a9284bc-kube-api-access-dwmfg" (OuterVolumeSpecName: "kube-api-access-dwmfg") pod "ea062396-352c-4a83-8745-5bec1a9284bc" (UID: "ea062396-352c-4a83-8745-5bec1a9284bc"). InnerVolumeSpecName "kube-api-access-dwmfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.430768 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a0c2c44-843d-4750-abbe-b4a9effe469f-kube-api-access-5rqrt" (OuterVolumeSpecName: "kube-api-access-5rqrt") pod "3a0c2c44-843d-4750-abbe-b4a9effe469f" (UID: "3a0c2c44-843d-4750-abbe-b4a9effe469f"). InnerVolumeSpecName "kube-api-access-5rqrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.462605 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.513029 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989f07ae-40ac-4b8d-91de-167df369b845-operator-scripts\") pod \"989f07ae-40ac-4b8d-91de-167df369b845\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.513094 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-operator-scripts\") pod \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.513219 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb7q7\" (UniqueName: \"kubernetes.io/projected/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-kube-api-access-tb7q7\") pod \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\" (UID: \"bfbaf5e7-e351-4c00-83e1-a3131589e9d9\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.513430 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6j7q\" (UniqueName: \"kubernetes.io/projected/989f07ae-40ac-4b8d-91de-167df369b845-kube-api-access-p6j7q\") pod \"989f07ae-40ac-4b8d-91de-167df369b845\" (UID: \"989f07ae-40ac-4b8d-91de-167df369b845\") " Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.513854 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/989f07ae-40ac-4b8d-91de-167df369b845-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "989f07ae-40ac-4b8d-91de-167df369b845" (UID: "989f07ae-40ac-4b8d-91de-167df369b845"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.513904 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bfbaf5e7-e351-4c00-83e1-a3131589e9d9" (UID: "bfbaf5e7-e351-4c00-83e1-a3131589e9d9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514217 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea062396-352c-4a83-8745-5bec1a9284bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514236 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989f07ae-40ac-4b8d-91de-167df369b845-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514245 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514255 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbkwp\" (UniqueName: \"kubernetes.io/projected/b071b469-ce94-4165-a933-55f9f8f36abe-kube-api-access-vbkwp\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514264 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rqrt\" (UniqueName: \"kubernetes.io/projected/3a0c2c44-843d-4750-abbe-b4a9effe469f-kube-api-access-5rqrt\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514273 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8tkg\" (UniqueName: \"kubernetes.io/projected/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-kube-api-access-q8tkg\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514281 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514290 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwmfg\" (UniqueName: \"kubernetes.io/projected/ea062396-352c-4a83-8745-5bec1a9284bc-kube-api-access-dwmfg\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514297 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a0c2c44-843d-4750-abbe-b4a9effe469f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.514306 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b071b469-ce94-4165-a933-55f9f8f36abe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.517339 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-kube-api-access-tb7q7" (OuterVolumeSpecName: "kube-api-access-tb7q7") pod "bfbaf5e7-e351-4c00-83e1-a3131589e9d9" (UID: "bfbaf5e7-e351-4c00-83e1-a3131589e9d9"). InnerVolumeSpecName "kube-api-access-tb7q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.519160 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989f07ae-40ac-4b8d-91de-167df369b845-kube-api-access-p6j7q" (OuterVolumeSpecName: "kube-api-access-p6j7q") pod "989f07ae-40ac-4b8d-91de-167df369b845" (UID: "989f07ae-40ac-4b8d-91de-167df369b845"). InnerVolumeSpecName "kube-api-access-p6j7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.615345 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6j7q\" (UniqueName: \"kubernetes.io/projected/989f07ae-40ac-4b8d-91de-167df369b845-kube-api-access-p6j7q\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.615388 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb7q7\" (UniqueName: \"kubernetes.io/projected/bfbaf5e7-e351-4c00-83e1-a3131589e9d9-kube-api-access-tb7q7\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.740422 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-41f6-account-create-update-ctjs8" event={"ID":"3a0c2c44-843d-4750-abbe-b4a9effe469f","Type":"ContainerDied","Data":"2b1fd4519b10be5b5808f75f16d2a4f54a0dc3cdec658767958bb4e135b57166"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.740484 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b1fd4519b10be5b5808f75f16d2a4f54a0dc3cdec658767958bb4e135b57166" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.740483 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-41f6-account-create-update-ctjs8" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.744602 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tmptx" event={"ID":"989f07ae-40ac-4b8d-91de-167df369b845","Type":"ContainerDied","Data":"c8acecd5105f329c79e8bba24f0d7621425900766ea61455668a6c506068bd92"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.744738 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8acecd5105f329c79e8bba24f0d7621425900766ea61455668a6c506068bd92" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.744825 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tmptx" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.765654 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wlwr5" event={"ID":"bd467839-0575-42ac-98fd-4e6d2cb3ceba","Type":"ContainerStarted","Data":"f32dd0f8de4ed06f6817844bfac86477a312740b57b1629d864513823a311230"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.770243 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v6sh8" event={"ID":"b071b469-ce94-4165-a933-55f9f8f36abe","Type":"ContainerDied","Data":"a1e667d6c16d50648de895728817216cb37996d3d472fb66b2603e40268dffed"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.770561 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1e667d6c16d50648de895728817216cb37996d3d472fb66b2603e40268dffed" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.770956 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v6sh8" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.778509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5b79-account-create-update-bdgk2" event={"ID":"dabcb5c7-61d1-46eb-9c15-c14a1e0df79e","Type":"ContainerDied","Data":"b1a2dbfb0c8e0315c57d84bef8424c27bb21c5fd60752e5f3714c967bede54af"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.781109 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1a2dbfb0c8e0315c57d84bef8424c27bb21c5fd60752e5f3714c967bede54af" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.778605 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5b79-account-create-update-bdgk2" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.787650 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sh2ng" event={"ID":"bfbaf5e7-e351-4c00-83e1-a3131589e9d9","Type":"ContainerDied","Data":"36ef7347921ee69c311aae113c461b89b0087623dd419c8938e4033ee0317836"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.787739 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36ef7347921ee69c311aae113c461b89b0087623dd419c8938e4033ee0317836" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.788188 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sh2ng" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.806137 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f5f0-account-create-update-gl7hp" event={"ID":"ea062396-352c-4a83-8745-5bec1a9284bc","Type":"ContainerDied","Data":"d1a395127a3487a29d8040e62b2f5b4fbcdaacd28c67c213fe9f49f4ea04ebd0"} Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.806494 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a395127a3487a29d8040e62b2f5b4fbcdaacd28c67c213fe9f49f4ea04ebd0" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.806590 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f5f0-account-create-update-gl7hp" Dec 04 06:29:37 crc kubenswrapper[4685]: I1204 06:29:37.829034 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wlwr5" podStartSLOduration=4.838771561 podStartE2EDuration="9.828991992s" podCreationTimestamp="2025-12-04 06:29:28 +0000 UTC" firstStartedPulling="2025-12-04 06:29:32.255252812 +0000 UTC m=+1109.443483627" lastFinishedPulling="2025-12-04 06:29:37.245473283 +0000 UTC m=+1114.433704058" observedRunningTime="2025-12-04 06:29:37.796339175 +0000 UTC m=+1114.984569960" watchObservedRunningTime="2025-12-04 06:29:37.828991992 +0000 UTC m=+1115.017222797" Dec 04 06:29:40 crc kubenswrapper[4685]: I1204 06:29:40.840328 4685 generic.go:334] "Generic (PLEG): container finished" podID="bd467839-0575-42ac-98fd-4e6d2cb3ceba" containerID="f32dd0f8de4ed06f6817844bfac86477a312740b57b1629d864513823a311230" exitCode=0 Dec 04 06:29:40 crc kubenswrapper[4685]: I1204 06:29:40.840473 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wlwr5" event={"ID":"bd467839-0575-42ac-98fd-4e6d2cb3ceba","Type":"ContainerDied","Data":"f32dd0f8de4ed06f6817844bfac86477a312740b57b1629d864513823a311230"} Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.232114 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.307385 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2sp8\" (UniqueName: \"kubernetes.io/projected/bd467839-0575-42ac-98fd-4e6d2cb3ceba-kube-api-access-v2sp8\") pod \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.307526 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-combined-ca-bundle\") pod \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.307554 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-config-data\") pod \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\" (UID: \"bd467839-0575-42ac-98fd-4e6d2cb3ceba\") " Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.319134 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd467839-0575-42ac-98fd-4e6d2cb3ceba-kube-api-access-v2sp8" (OuterVolumeSpecName: "kube-api-access-v2sp8") pod "bd467839-0575-42ac-98fd-4e6d2cb3ceba" (UID: "bd467839-0575-42ac-98fd-4e6d2cb3ceba"). InnerVolumeSpecName "kube-api-access-v2sp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.338703 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd467839-0575-42ac-98fd-4e6d2cb3ceba" (UID: "bd467839-0575-42ac-98fd-4e6d2cb3ceba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.365327 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-config-data" (OuterVolumeSpecName: "config-data") pod "bd467839-0575-42ac-98fd-4e6d2cb3ceba" (UID: "bd467839-0575-42ac-98fd-4e6d2cb3ceba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.410296 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.410338 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2sp8\" (UniqueName: \"kubernetes.io/projected/bd467839-0575-42ac-98fd-4e6d2cb3ceba-kube-api-access-v2sp8\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.410353 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd467839-0575-42ac-98fd-4e6d2cb3ceba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.867865 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wlwr5" event={"ID":"bd467839-0575-42ac-98fd-4e6d2cb3ceba","Type":"ContainerDied","Data":"35d8ccf908e59c3de1846f3af8c8d84efeaba4612adf09b224bbd30a0ee6b80d"} Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.868348 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35d8ccf908e59c3de1846f3af8c8d84efeaba4612adf09b224bbd30a0ee6b80d" Dec 04 06:29:42 crc kubenswrapper[4685]: I1204 06:29:42.868164 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wlwr5" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.238560 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nkpwl"] Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242109 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea062396-352c-4a83-8745-5bec1a9284bc" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242140 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea062396-352c-4a83-8745-5bec1a9284bc" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242151 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242158 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242181 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd467839-0575-42ac-98fd-4e6d2cb3ceba" containerName="keystone-db-sync" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242189 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd467839-0575-42ac-98fd-4e6d2cb3ceba" containerName="keystone-db-sync" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242206 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbaf5e7-e351-4c00-83e1-a3131589e9d9" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242212 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbaf5e7-e351-4c00-83e1-a3131589e9d9" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242234 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerName="dnsmasq-dns" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242240 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerName="dnsmasq-dns" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242263 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989f07ae-40ac-4b8d-91de-167df369b845" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242269 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="989f07ae-40ac-4b8d-91de-167df369b845" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242276 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerName="init" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242282 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerName="init" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242300 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b071b469-ce94-4165-a933-55f9f8f36abe" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242306 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b071b469-ce94-4165-a933-55f9f8f36abe" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: E1204 06:29:43.242317 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0c2c44-843d-4750-abbe-b4a9effe469f" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242338 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0c2c44-843d-4750-abbe-b4a9effe469f" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242549 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4c5c56-929c-47b1-a488-f9eea7edeccc" containerName="dnsmasq-dns" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242599 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242611 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd467839-0575-42ac-98fd-4e6d2cb3ceba" containerName="keystone-db-sync" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242619 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfbaf5e7-e351-4c00-83e1-a3131589e9d9" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242628 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea062396-352c-4a83-8745-5bec1a9284bc" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242638 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b071b469-ce94-4165-a933-55f9f8f36abe" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242666 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="989f07ae-40ac-4b8d-91de-167df369b845" containerName="mariadb-database-create" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.242677 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0c2c44-843d-4750-abbe-b4a9effe469f" containerName="mariadb-account-create-update" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.243458 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.247888 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.248154 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.248611 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.248764 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c6l2p" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.248868 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.254513 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wxsxd"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.258101 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.265400 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wxsxd"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.284962 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nkpwl"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.338925 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-combined-ca-bundle\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339246 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339345 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-credential-keys\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339454 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-scripts\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339543 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kln2b\" (UniqueName: \"kubernetes.io/projected/80875460-dbe1-4b1e-80cf-20c7484dd40b-kube-api-access-kln2b\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339796 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339861 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.339940 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-config-data\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.340009 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp4n8\" (UniqueName: \"kubernetes.io/projected/25dfcb44-a95f-466e-b9fe-f8080cea09e8-kube-api-access-lp4n8\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.340057 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-fernet-keys\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.340165 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.340231 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-config\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.391797 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5588ffff99-v5xfx"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.393206 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.395441 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.395786 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.400561 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-z5xkm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.400816 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.421954 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5588ffff99-v5xfx"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460081 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-config-data\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460160 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp4n8\" (UniqueName: \"kubernetes.io/projected/25dfcb44-a95f-466e-b9fe-f8080cea09e8-kube-api-access-lp4n8\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460203 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-fernet-keys\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460240 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460265 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgkrf\" (UniqueName: \"kubernetes.io/projected/5fa411a1-1dc0-4140-b7f7-8684304402a5-kube-api-access-jgkrf\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460301 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-config\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460338 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-combined-ca-bundle\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460364 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5fa411a1-1dc0-4140-b7f7-8684304402a5-horizon-secret-key\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460398 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460444 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-credential-keys\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460462 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-config-data\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460491 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-scripts\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460524 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-scripts\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460560 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kln2b\" (UniqueName: \"kubernetes.io/projected/80875460-dbe1-4b1e-80cf-20c7484dd40b-kube-api-access-kln2b\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460597 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460618 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.460643 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fa411a1-1dc0-4140-b7f7-8684304402a5-logs\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.462074 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.463534 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.463784 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.464117 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.464380 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-config\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.464977 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-bcqrm"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.472220 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.483265 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-p9tjg" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.483591 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.483740 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.517274 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp4n8\" (UniqueName: \"kubernetes.io/projected/25dfcb44-a95f-466e-b9fe-f8080cea09e8-kube-api-access-lp4n8\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.525316 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bcqrm"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.535042 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-credential-keys\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.535344 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-scripts\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.538108 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-fernet-keys\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.541185 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-config-data\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.542843 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-combined-ca-bundle\") pod \"keystone-bootstrap-nkpwl\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.550547 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.557801 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.564702 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.565198 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.566902 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fa411a1-1dc0-4140-b7f7-8684304402a5-logs\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567076 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxl4\" (UniqueName: \"kubernetes.io/projected/4ab9f58d-e6bb-445d-a748-0059a8981f49-kube-api-access-8qxl4\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567184 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgkrf\" (UniqueName: \"kubernetes.io/projected/5fa411a1-1dc0-4140-b7f7-8684304402a5-kube-api-access-jgkrf\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567273 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-config\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567382 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5fa411a1-1dc0-4140-b7f7-8684304402a5-horizon-secret-key\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567526 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-combined-ca-bundle\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567616 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-config-data\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567719 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-scripts\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.567766 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fa411a1-1dc0-4140-b7f7-8684304402a5-logs\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.568951 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-config-data\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.569987 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-scripts\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.570141 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kln2b\" (UniqueName: \"kubernetes.io/projected/80875460-dbe1-4b1e-80cf-20c7484dd40b-kube-api-access-kln2b\") pod \"dnsmasq-dns-bbf5cc879-wxsxd\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.571890 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.575189 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5fa411a1-1dc0-4140-b7f7-8684304402a5-horizon-secret-key\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.631657 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.633733 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgkrf\" (UniqueName: \"kubernetes.io/projected/5fa411a1-1dc0-4140-b7f7-8684304402a5-kube-api-access-jgkrf\") pod \"horizon-5588ffff99-v5xfx\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.633892 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.669205 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.669789 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86lzg\" (UniqueName: \"kubernetes.io/projected/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-kube-api-access-86lzg\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.669911 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-run-httpd\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.670011 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-config-data\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.670123 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-log-httpd\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.670235 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxl4\" (UniqueName: \"kubernetes.io/projected/4ab9f58d-e6bb-445d-a748-0059a8981f49-kube-api-access-8qxl4\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.670366 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-scripts\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.670508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.670786 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-config\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.676044 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-combined-ca-bundle\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.687529 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-config\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.688464 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-combined-ca-bundle\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.695763 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5498d"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.697091 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.705704 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-p7d29" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.705935 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.708000 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5498d"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.710465 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.721363 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxl4\" (UniqueName: \"kubernetes.io/projected/4ab9f58d-e6bb-445d-a748-0059a8981f49-kube-api-access-8qxl4\") pod \"neutron-db-sync-bcqrm\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.740798 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.778776 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86lzg\" (UniqueName: \"kubernetes.io/projected/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-kube-api-access-86lzg\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779072 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-run-httpd\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779156 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-config-data\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779246 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-log-httpd\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779351 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-scripts\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779460 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779609 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-db-sync-config-data\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.791638 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48pl6\" (UniqueName: \"kubernetes.io/projected/6f6bce91-0f30-4543-ac33-b34877206b05-kube-api-access-48pl6\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.791784 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-combined-ca-bundle\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.779813 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.780202 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-log-httpd\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.780824 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-run-httpd\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.793088 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.808582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-config-data\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.809542 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.810784 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.814941 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-scripts\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.816001 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86lzg\" (UniqueName: \"kubernetes.io/projected/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-kube-api-access-86lzg\") pod \"ceilometer-0\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " pod="openstack/ceilometer-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.819206 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6fb76c6c8c-j6b94"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.820307 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.820808 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.827168 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-66ml8" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.828506 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.828637 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.828743 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.891281 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894422 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdxjl\" (UniqueName: \"kubernetes.io/projected/f908631b-9f25-474e-9dff-815a7dc11aae-kube-api-access-pdxjl\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894457 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-464vx\" (UniqueName: \"kubernetes.io/projected/6d9a855f-3727-41c7-85a9-d404efbd4d83-kube-api-access-464vx\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894509 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894533 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d9a855f-3727-41c7-85a9-d404efbd4d83-horizon-secret-key\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894566 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894591 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d9a855f-3727-41c7-85a9-d404efbd4d83-logs\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894645 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-scripts\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894673 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-config-data\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894702 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-db-sync-config-data\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894724 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894751 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-scripts\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894774 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-config-data\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894795 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48pl6\" (UniqueName: \"kubernetes.io/projected/6f6bce91-0f30-4543-ac33-b34877206b05-kube-api-access-48pl6\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894816 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894844 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-combined-ca-bundle\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.894865 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-logs\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.908070 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-combined-ca-bundle\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.912910 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-db-sync-config-data\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.918226 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48pl6\" (UniqueName: \"kubernetes.io/projected/6f6bce91-0f30-4543-ac33-b34877206b05-kube-api-access-48pl6\") pod \"barbican-db-sync-5498d\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.935248 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fb76c6c8c-j6b94"] Dec 04 06:29:43 crc kubenswrapper[4685]: I1204 06:29:43.994134 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cpkd7"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.000613 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.002877 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.002931 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d9a855f-3727-41c7-85a9-d404efbd4d83-horizon-secret-key\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.002959 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.002982 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d9a855f-3727-41c7-85a9-d404efbd4d83-logs\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003021 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-scripts\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003046 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-config-data\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003075 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-scripts\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003120 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-config-data\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003145 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003176 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-logs\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003204 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdxjl\" (UniqueName: \"kubernetes.io/projected/f908631b-9f25-474e-9dff-815a7dc11aae-kube-api-access-pdxjl\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.003224 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-464vx\" (UniqueName: \"kubernetes.io/projected/6d9a855f-3727-41c7-85a9-d404efbd4d83-kube-api-access-464vx\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.004604 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ccdtn" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.004788 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.005061 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.008209 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.008246 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d9a855f-3727-41c7-85a9-d404efbd4d83-logs\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.008823 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-scripts\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.011194 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-config-data\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.012753 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.017528 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cpkd7"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.024325 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-logs\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.024747 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d9a855f-3727-41c7-85a9-d404efbd4d83-horizon-secret-key\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.056986 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-scripts\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.057114 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wxsxd"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.057656 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.062755 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.067017 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.070089 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-config-data\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.072446 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-464vx\" (UniqueName: \"kubernetes.io/projected/6d9a855f-3727-41c7-85a9-d404efbd4d83-kube-api-access-464vx\") pod \"horizon-6fb76c6c8c-j6b94\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.074213 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdxjl\" (UniqueName: \"kubernetes.io/projected/f908631b-9f25-474e-9dff-815a7dc11aae-kube-api-access-pdxjl\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.075270 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.085999 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5498d" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.091778 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-59cdh"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.093211 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.096135 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.096465 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.096690 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4nxjd" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.104988 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-config-data\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.105168 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-combined-ca-bundle\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.105305 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf9kv\" (UniqueName: \"kubernetes.io/projected/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-kube-api-access-rf9kv\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.105456 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-logs\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.105590 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-scripts\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.133990 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-59cdh"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.148465 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-xd9fs"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.151990 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.163160 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-xd9fs"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.182456 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.187449 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.187752 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.196219 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.197972 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208240 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx8tj\" (UniqueName: \"kubernetes.io/projected/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-kube-api-access-gx8tj\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208289 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208313 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf9kv\" (UniqueName: \"kubernetes.io/projected/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-kube-api-access-rf9kv\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208333 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-config\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208364 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208428 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-logs\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-scripts\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208506 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-config-data\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208532 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208557 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.208583 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-combined-ca-bundle\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.217530 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-combined-ca-bundle\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.217537 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-config-data\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.217694 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-logs\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.221922 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-scripts\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.229804 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf9kv\" (UniqueName: \"kubernetes.io/projected/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-kube-api-access-rf9kv\") pod \"placement-db-sync-cpkd7\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.246067 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.283832 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315597 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315694 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315724 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315749 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315809 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-db-sync-config-data\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315844 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-scripts\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315879 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssnzm\" (UniqueName: \"kubernetes.io/projected/98de6ea0-e984-4319-a585-d70e40f4291f-kube-api-access-ssnzm\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315909 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx8tj\" (UniqueName: \"kubernetes.io/projected/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-kube-api-access-gx8tj\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315965 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.315996 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-config\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316054 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316079 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-config-data\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316261 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-logs\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316434 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316470 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316499 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/964316c7-6da5-4725-8140-6d249ee01a68-etc-machine-id\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316547 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-combined-ca-bundle\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316649 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316710 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psn2s\" (UniqueName: \"kubernetes.io/projected/964316c7-6da5-4725-8140-6d249ee01a68-kube-api-access-psn2s\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.316758 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.317584 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.317648 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.318230 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-config\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.323537 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.326294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.327306 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wxsxd"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.338630 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cpkd7" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.344458 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx8tj\" (UniqueName: \"kubernetes.io/projected/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-kube-api-access-gx8tj\") pod \"dnsmasq-dns-56df8fb6b7-xd9fs\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418275 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psn2s\" (UniqueName: \"kubernetes.io/projected/964316c7-6da5-4725-8140-6d249ee01a68-kube-api-access-psn2s\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418504 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418546 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418581 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418623 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-db-sync-config-data\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418652 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-scripts\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418682 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssnzm\" (UniqueName: \"kubernetes.io/projected/98de6ea0-e984-4319-a585-d70e40f4291f-kube-api-access-ssnzm\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418747 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-config-data\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418786 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-logs\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418841 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418882 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/964316c7-6da5-4725-8140-6d249ee01a68-etc-machine-id\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418906 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418929 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-combined-ca-bundle\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.418980 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.419609 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-logs\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.421164 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.422277 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.422542 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.422680 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/964316c7-6da5-4725-8140-6d249ee01a68-etc-machine-id\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.428014 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.429081 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.439549 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-db-sync-config-data\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.439547 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.439663 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-scripts\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.439692 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-combined-ca-bundle\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.440087 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-config-data\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.440448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psn2s\" (UniqueName: \"kubernetes.io/projected/964316c7-6da5-4725-8140-6d249ee01a68-kube-api-access-psn2s\") pod \"cinder-db-sync-59cdh\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.456006 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nkpwl"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.457436 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssnzm\" (UniqueName: \"kubernetes.io/projected/98de6ea0-e984-4319-a585-d70e40f4291f-kube-api-access-ssnzm\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.478516 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.552214 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.588781 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bcqrm"] Dec 04 06:29:44 crc kubenswrapper[4685]: W1204 06:29:44.590635 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fa411a1_1dc0_4140_b7f7_8684304402a5.slice/crio-06bda29bf5c9e409e287ad60f9e365248b522bfdd44fa040a8b10b778986f22e WatchSource:0}: Error finding container 06bda29bf5c9e409e287ad60f9e365248b522bfdd44fa040a8b10b778986f22e: Status 404 returned error can't find the container with id 06bda29bf5c9e409e287ad60f9e365248b522bfdd44fa040a8b10b778986f22e Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.611316 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5588ffff99-v5xfx"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.739385 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-59cdh" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.825995 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.838320 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.921933 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcqrm" event={"ID":"4ab9f58d-e6bb-445d-a748-0059a8981f49","Type":"ContainerStarted","Data":"066a7793eec3e8427e39f0770c8ff919fbd93b10dd22c3f4c4a3320545ba35df"} Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.926983 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" event={"ID":"80875460-dbe1-4b1e-80cf-20c7484dd40b","Type":"ContainerStarted","Data":"cc094ca9b87443e25abd1edf38717cba4320de168963c7b7cf50ddebfbf9704b"} Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.929471 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkpwl" event={"ID":"25dfcb44-a95f-466e-b9fe-f8080cea09e8","Type":"ContainerStarted","Data":"f27ec6355626100c05d9ec57e354ceddd4ab4fb43c70e1d3797a2d9103ea75a4"} Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.930739 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerStarted","Data":"20652d1091ddaeab7d547e06db60c052121ee0cea8c3d1ab5e379a90eadc5b13"} Dec 04 06:29:44 crc kubenswrapper[4685]: I1204 06:29:44.931902 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5588ffff99-v5xfx" event={"ID":"5fa411a1-1dc0-4140-b7f7-8684304402a5","Type":"ContainerStarted","Data":"06bda29bf5c9e409e287ad60f9e365248b522bfdd44fa040a8b10b778986f22e"} Dec 04 06:29:45 crc kubenswrapper[4685]: W1204 06:29:45.197023 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c76ec1c_6d8c_48e0_9128_4fd9fb3cbed3.slice/crio-5bd8e4476691b03027da40d0743abc8e8b9f857c3f6626d33332a7e84189d72a WatchSource:0}: Error finding container 5bd8e4476691b03027da40d0743abc8e8b9f857c3f6626d33332a7e84189d72a: Status 404 returned error can't find the container with id 5bd8e4476691b03027da40d0743abc8e8b9f857c3f6626d33332a7e84189d72a Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.198955 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cpkd7"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.217936 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5498d"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.221784 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fb76c6c8c-j6b94"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.285261 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-59cdh"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.310937 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-xd9fs"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.339558 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.611177 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.717667 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.724542 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5588ffff99-v5xfx"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.826549 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6ddd9d7c47-dwpfz"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.828096 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.848890 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.891032 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b56721cb-e8ce-4106-920b-0b735ef29a38-horizon-secret-key\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.891128 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-scripts\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.891214 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-config-data\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.891243 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2dzb\" (UniqueName: \"kubernetes.io/projected/b56721cb-e8ce-4106-920b-0b735ef29a38-kube-api-access-f2dzb\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.891268 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b56721cb-e8ce-4106-920b-0b735ef29a38-logs\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.891455 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ddd9d7c47-dwpfz"] Dec 04 06:29:45 crc kubenswrapper[4685]: I1204 06:29:45.989749 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98de6ea0-e984-4319-a585-d70e40f4291f","Type":"ContainerStarted","Data":"b97bd201ac1e78892db145b586d5065803b23400ff62e05ad36bd45865c83c8f"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.010750 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b56721cb-e8ce-4106-920b-0b735ef29a38-horizon-secret-key\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.010828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-scripts\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.010913 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-config-data\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.010938 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2dzb\" (UniqueName: \"kubernetes.io/projected/b56721cb-e8ce-4106-920b-0b735ef29a38-kube-api-access-f2dzb\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.010959 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b56721cb-e8ce-4106-920b-0b735ef29a38-logs\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.011506 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b56721cb-e8ce-4106-920b-0b735ef29a38-logs\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.037740 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b56721cb-e8ce-4106-920b-0b735ef29a38-horizon-secret-key\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.038183 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-scripts\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.039108 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-config-data\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.045812 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f908631b-9f25-474e-9dff-815a7dc11aae","Type":"ContainerStarted","Data":"ba797b8fe42614311931fca52ba0a5ac490d81901c8b947bb83fe6159edfe41a"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.077888 4685 generic.go:334] "Generic (PLEG): container finished" podID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerID="dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486" exitCode=0 Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.077968 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" event={"ID":"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7","Type":"ContainerDied","Data":"dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.078000 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" event={"ID":"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7","Type":"ContainerStarted","Data":"355451359a7464821e403bca6e4cb1703fdc5401d7563269c57f16efce0e4dda"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.121385 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cpkd7" event={"ID":"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3","Type":"ContainerStarted","Data":"5bd8e4476691b03027da40d0743abc8e8b9f857c3f6626d33332a7e84189d72a"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.149054 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2dzb\" (UniqueName: \"kubernetes.io/projected/b56721cb-e8ce-4106-920b-0b735ef29a38-kube-api-access-f2dzb\") pod \"horizon-6ddd9d7c47-dwpfz\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.159709 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.169364 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5498d" event={"ID":"6f6bce91-0f30-4543-ac33-b34877206b05","Type":"ContainerStarted","Data":"6611159eddc7c9c87ead02365070f6389b0e9507202bb41edb1ff7681210b158"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.172113 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-59cdh" event={"ID":"964316c7-6da5-4725-8140-6d249ee01a68","Type":"ContainerStarted","Data":"a239456b7df1c8a7bc0119a81e155fe9044fe10c2954e7d31b8149eccb6b244f"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.181439 4685 generic.go:334] "Generic (PLEG): container finished" podID="80875460-dbe1-4b1e-80cf-20c7484dd40b" containerID="b4e5f2d540af3385117ee8c848b29ad1c92c72760f7466a80049c05a775a6bf0" exitCode=0 Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.181545 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" event={"ID":"80875460-dbe1-4b1e-80cf-20c7484dd40b","Type":"ContainerDied","Data":"b4e5f2d540af3385117ee8c848b29ad1c92c72760f7466a80049c05a775a6bf0"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.186245 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb76c6c8c-j6b94" event={"ID":"6d9a855f-3727-41c7-85a9-d404efbd4d83","Type":"ContainerStarted","Data":"92a6ab637515273894234fa0a3bd94d01d819ece1085e1c30bdd6805eb6e74b3"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.257220 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.289869 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkpwl" event={"ID":"25dfcb44-a95f-466e-b9fe-f8080cea09e8","Type":"ContainerStarted","Data":"5e32e6824d22ae1aa3272b6cd47136721ab3ad5274f2f968f1bc3abb29d018c5"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.293121 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcqrm" event={"ID":"4ab9f58d-e6bb-445d-a748-0059a8981f49","Type":"ContainerStarted","Data":"9472458b1f7b017c0ddb1d80493683f3e4e415e4cdc6387da6cd9b1b21bd7729"} Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.308857 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nkpwl" podStartSLOduration=3.308837549 podStartE2EDuration="3.308837549s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:46.304586636 +0000 UTC m=+1123.492817411" watchObservedRunningTime="2025-12-04 06:29:46.308837549 +0000 UTC m=+1123.497068324" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.334070 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-bcqrm" podStartSLOduration=3.334048782 podStartE2EDuration="3.334048782s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:46.330889684 +0000 UTC m=+1123.519120459" watchObservedRunningTime="2025-12-04 06:29:46.334048782 +0000 UTC m=+1123.522279557" Dec 04 06:29:46 crc kubenswrapper[4685]: I1204 06:29:46.901978 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ddd9d7c47-dwpfz"] Dec 04 06:29:46 crc kubenswrapper[4685]: W1204 06:29:46.926944 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb56721cb_e8ce_4106_920b_0b735ef29a38.slice/crio-99a13dcde48e5e5ae7f7a6500673b9221155b2f40d39f1e55b7ad7bf161fd97c WatchSource:0}: Error finding container 99a13dcde48e5e5ae7f7a6500673b9221155b2f40d39f1e55b7ad7bf161fd97c: Status 404 returned error can't find the container with id 99a13dcde48e5e5ae7f7a6500673b9221155b2f40d39f1e55b7ad7bf161fd97c Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.070371 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.248650 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-swift-storage-0\") pod \"80875460-dbe1-4b1e-80cf-20c7484dd40b\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.248706 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-config\") pod \"80875460-dbe1-4b1e-80cf-20c7484dd40b\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.248853 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kln2b\" (UniqueName: \"kubernetes.io/projected/80875460-dbe1-4b1e-80cf-20c7484dd40b-kube-api-access-kln2b\") pod \"80875460-dbe1-4b1e-80cf-20c7484dd40b\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.248888 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-sb\") pod \"80875460-dbe1-4b1e-80cf-20c7484dd40b\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.248959 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-svc\") pod \"80875460-dbe1-4b1e-80cf-20c7484dd40b\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.248982 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-nb\") pod \"80875460-dbe1-4b1e-80cf-20c7484dd40b\" (UID: \"80875460-dbe1-4b1e-80cf-20c7484dd40b\") " Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.265645 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80875460-dbe1-4b1e-80cf-20c7484dd40b-kube-api-access-kln2b" (OuterVolumeSpecName: "kube-api-access-kln2b") pod "80875460-dbe1-4b1e-80cf-20c7484dd40b" (UID: "80875460-dbe1-4b1e-80cf-20c7484dd40b"). InnerVolumeSpecName "kube-api-access-kln2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.280332 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "80875460-dbe1-4b1e-80cf-20c7484dd40b" (UID: "80875460-dbe1-4b1e-80cf-20c7484dd40b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.291857 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "80875460-dbe1-4b1e-80cf-20c7484dd40b" (UID: "80875460-dbe1-4b1e-80cf-20c7484dd40b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.305885 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-config" (OuterVolumeSpecName: "config") pod "80875460-dbe1-4b1e-80cf-20c7484dd40b" (UID: "80875460-dbe1-4b1e-80cf-20c7484dd40b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.306733 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "80875460-dbe1-4b1e-80cf-20c7484dd40b" (UID: "80875460-dbe1-4b1e-80cf-20c7484dd40b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.307518 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "80875460-dbe1-4b1e-80cf-20c7484dd40b" (UID: "80875460-dbe1-4b1e-80cf-20c7484dd40b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.313450 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.313459 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wxsxd" event={"ID":"80875460-dbe1-4b1e-80cf-20c7484dd40b","Type":"ContainerDied","Data":"cc094ca9b87443e25abd1edf38717cba4320de168963c7b7cf50ddebfbf9704b"} Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.313589 4685 scope.go:117] "RemoveContainer" containerID="b4e5f2d540af3385117ee8c848b29ad1c92c72760f7466a80049c05a775a6bf0" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.321916 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddd9d7c47-dwpfz" event={"ID":"b56721cb-e8ce-4106-920b-0b735ef29a38","Type":"ContainerStarted","Data":"99a13dcde48e5e5ae7f7a6500673b9221155b2f40d39f1e55b7ad7bf161fd97c"} Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.351805 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.351856 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.351887 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.351900 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.351913 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80875460-dbe1-4b1e-80cf-20c7484dd40b-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.351926 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kln2b\" (UniqueName: \"kubernetes.io/projected/80875460-dbe1-4b1e-80cf-20c7484dd40b-kube-api-access-kln2b\") on node \"crc\" DevicePath \"\"" Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.441236 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wxsxd"] Dec 04 06:29:47 crc kubenswrapper[4685]: I1204 06:29:47.460995 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wxsxd"] Dec 04 06:29:48 crc kubenswrapper[4685]: I1204 06:29:48.348467 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f908631b-9f25-474e-9dff-815a7dc11aae","Type":"ContainerStarted","Data":"6e0ad47fe1e6118f1a881b2cc0b01072646fbbc4f857be28eff12ec66a93a635"} Dec 04 06:29:48 crc kubenswrapper[4685]: I1204 06:29:48.352372 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" event={"ID":"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7","Type":"ContainerStarted","Data":"08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b"} Dec 04 06:29:48 crc kubenswrapper[4685]: I1204 06:29:48.352465 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:48 crc kubenswrapper[4685]: I1204 06:29:48.362725 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98de6ea0-e984-4319-a585-d70e40f4291f","Type":"ContainerStarted","Data":"c912f3f748574799ef519291dc29aa36647a8cee1888485f19e13ccdead9b27d"} Dec 04 06:29:48 crc kubenswrapper[4685]: I1204 06:29:48.382311 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" podStartSLOduration=5.382290188 podStartE2EDuration="5.382290188s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:48.381225805 +0000 UTC m=+1125.569456600" watchObservedRunningTime="2025-12-04 06:29:48.382290188 +0000 UTC m=+1125.570520963" Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.142053 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80875460-dbe1-4b1e-80cf-20c7484dd40b" path="/var/lib/kubelet/pods/80875460-dbe1-4b1e-80cf-20c7484dd40b/volumes" Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.385088 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f908631b-9f25-474e-9dff-815a7dc11aae","Type":"ContainerStarted","Data":"1f05d49f2d3688095750e25d589e7541d9586496e29a620e646130ff778717c0"} Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.385303 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-log" containerID="cri-o://6e0ad47fe1e6118f1a881b2cc0b01072646fbbc4f857be28eff12ec66a93a635" gracePeriod=30 Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.385967 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-httpd" containerID="cri-o://1f05d49f2d3688095750e25d589e7541d9586496e29a620e646130ff778717c0" gracePeriod=30 Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.390822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98de6ea0-e984-4319-a585-d70e40f4291f","Type":"ContainerStarted","Data":"e5de6316c344d13be595a2691d616ff6e53470ddf3bbdac90998471d89c80a50"} Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.390814 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-log" containerID="cri-o://c912f3f748574799ef519291dc29aa36647a8cee1888485f19e13ccdead9b27d" gracePeriod=30 Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.391525 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-httpd" containerID="cri-o://e5de6316c344d13be595a2691d616ff6e53470ddf3bbdac90998471d89c80a50" gracePeriod=30 Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.415222 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.415175987 podStartE2EDuration="6.415175987s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:49.411760759 +0000 UTC m=+1126.599991534" watchObservedRunningTime="2025-12-04 06:29:49.415175987 +0000 UTC m=+1126.603406762" Dec 04 06:29:49 crc kubenswrapper[4685]: I1204 06:29:49.442942 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.44291738 podStartE2EDuration="6.44291738s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:29:49.4368962 +0000 UTC m=+1126.625126975" watchObservedRunningTime="2025-12-04 06:29:49.44291738 +0000 UTC m=+1126.631148155" Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.403907 4685 generic.go:334] "Generic (PLEG): container finished" podID="98de6ea0-e984-4319-a585-d70e40f4291f" containerID="e5de6316c344d13be595a2691d616ff6e53470ddf3bbdac90998471d89c80a50" exitCode=143 Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.404963 4685 generic.go:334] "Generic (PLEG): container finished" podID="98de6ea0-e984-4319-a585-d70e40f4291f" containerID="c912f3f748574799ef519291dc29aa36647a8cee1888485f19e13ccdead9b27d" exitCode=143 Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.403982 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98de6ea0-e984-4319-a585-d70e40f4291f","Type":"ContainerDied","Data":"e5de6316c344d13be595a2691d616ff6e53470ddf3bbdac90998471d89c80a50"} Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.405076 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98de6ea0-e984-4319-a585-d70e40f4291f","Type":"ContainerDied","Data":"c912f3f748574799ef519291dc29aa36647a8cee1888485f19e13ccdead9b27d"} Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.411667 4685 generic.go:334] "Generic (PLEG): container finished" podID="f908631b-9f25-474e-9dff-815a7dc11aae" containerID="1f05d49f2d3688095750e25d589e7541d9586496e29a620e646130ff778717c0" exitCode=0 Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.411706 4685 generic.go:334] "Generic (PLEG): container finished" podID="f908631b-9f25-474e-9dff-815a7dc11aae" containerID="6e0ad47fe1e6118f1a881b2cc0b01072646fbbc4f857be28eff12ec66a93a635" exitCode=143 Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.411737 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f908631b-9f25-474e-9dff-815a7dc11aae","Type":"ContainerDied","Data":"1f05d49f2d3688095750e25d589e7541d9586496e29a620e646130ff778717c0"} Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.411812 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f908631b-9f25-474e-9dff-815a7dc11aae","Type":"ContainerDied","Data":"6e0ad47fe1e6118f1a881b2cc0b01072646fbbc4f857be28eff12ec66a93a635"} Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.414489 4685 generic.go:334] "Generic (PLEG): container finished" podID="25dfcb44-a95f-466e-b9fe-f8080cea09e8" containerID="5e32e6824d22ae1aa3272b6cd47136721ab3ad5274f2f968f1bc3abb29d018c5" exitCode=0 Dec 04 06:29:50 crc kubenswrapper[4685]: I1204 06:29:50.414538 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkpwl" event={"ID":"25dfcb44-a95f-466e-b9fe-f8080cea09e8","Type":"ContainerDied","Data":"5e32e6824d22ae1aa3272b6cd47136721ab3ad5274f2f968f1bc3abb29d018c5"} Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.501521 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fb76c6c8c-j6b94"] Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.535707 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6bdfd7d6f6-8slln"] Dec 04 06:29:52 crc kubenswrapper[4685]: E1204 06:29:52.536245 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80875460-dbe1-4b1e-80cf-20c7484dd40b" containerName="init" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.536258 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="80875460-dbe1-4b1e-80cf-20c7484dd40b" containerName="init" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.536458 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="80875460-dbe1-4b1e-80cf-20c7484dd40b" containerName="init" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.537521 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.541030 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.545843 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bdfd7d6f6-8slln"] Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.584544 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8cjl\" (UniqueName: \"kubernetes.io/projected/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-kube-api-access-h8cjl\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.585391 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-secret-key\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.585441 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-logs\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.585512 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-config-data\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.585552 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-combined-ca-bundle\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.585573 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-scripts\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.585606 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-tls-certs\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.614540 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ddd9d7c47-dwpfz"] Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.654781 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c9548d9dd-h5vqf"] Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.658549 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.664937 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c9548d9dd-h5vqf"] Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686305 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-tls-certs\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686368 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8cjl\" (UniqueName: \"kubernetes.io/projected/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-kube-api-access-h8cjl\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686399 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-horizon-tls-certs\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686538 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-config-data\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686691 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-logs\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686766 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-secret-key\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686790 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-logs\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686816 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-scripts\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686850 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-horizon-secret-key\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686938 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-config-data\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.686987 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-combined-ca-bundle\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.687024 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-combined-ca-bundle\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.687046 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-scripts\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.687067 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgvbz\" (UniqueName: \"kubernetes.io/projected/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-kube-api-access-mgvbz\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.688173 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-logs\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.688547 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-config-data\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.689136 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-scripts\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.695395 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-secret-key\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.696509 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-tls-certs\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.708436 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8cjl\" (UniqueName: \"kubernetes.io/projected/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-kube-api-access-h8cjl\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.713596 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-combined-ca-bundle\") pod \"horizon-6bdfd7d6f6-8slln\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790060 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-combined-ca-bundle\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790136 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgvbz\" (UniqueName: \"kubernetes.io/projected/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-kube-api-access-mgvbz\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790227 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-horizon-tls-certs\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790277 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-config-data\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790330 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-logs\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790373 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-scripts\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.790402 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-horizon-secret-key\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.792122 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-logs\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.792915 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-scripts\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.793711 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-config-data\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.798949 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-combined-ca-bundle\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.799222 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-horizon-secret-key\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.806610 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-horizon-tls-certs\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.810782 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgvbz\" (UniqueName: \"kubernetes.io/projected/0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236-kube-api-access-mgvbz\") pod \"horizon-5c9548d9dd-h5vqf\" (UID: \"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236\") " pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:52 crc kubenswrapper[4685]: I1204 06:29:52.900845 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:29:53 crc kubenswrapper[4685]: I1204 06:29:53.077862 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:29:54 crc kubenswrapper[4685]: I1204 06:29:54.480858 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:29:54 crc kubenswrapper[4685]: I1204 06:29:54.559610 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9wqf7"] Dec 04 06:29:54 crc kubenswrapper[4685]: I1204 06:29:54.560837 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" containerID="cri-o://179ee4c4f3bb1d39611f2e3d12131ac9da868bf495deca3146a5b9cc40bd0ed6" gracePeriod=10 Dec 04 06:29:55 crc kubenswrapper[4685]: I1204 06:29:55.442597 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 04 06:29:55 crc kubenswrapper[4685]: I1204 06:29:55.499466 4685 generic.go:334] "Generic (PLEG): container finished" podID="991d5443-6323-48ba-9a55-832b4d80820b" containerID="179ee4c4f3bb1d39611f2e3d12131ac9da868bf495deca3146a5b9cc40bd0ed6" exitCode=0 Dec 04 06:29:55 crc kubenswrapper[4685]: I1204 06:29:55.499528 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" event={"ID":"991d5443-6323-48ba-9a55-832b4d80820b","Type":"ContainerDied","Data":"179ee4c4f3bb1d39611f2e3d12131ac9da868bf495deca3146a5b9cc40bd0ed6"} Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.158144 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4"] Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.160028 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.172152 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.172495 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.175147 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4"] Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.277833 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8daa431f-999a-4213-a745-9d8b03aa028e-secret-volume\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.277911 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8stp\" (UniqueName: \"kubernetes.io/projected/8daa431f-999a-4213-a745-9d8b03aa028e-kube-api-access-r8stp\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.278835 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8daa431f-999a-4213-a745-9d8b03aa028e-config-volume\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.380697 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8daa431f-999a-4213-a745-9d8b03aa028e-config-volume\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.380781 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8daa431f-999a-4213-a745-9d8b03aa028e-secret-volume\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.380828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8stp\" (UniqueName: \"kubernetes.io/projected/8daa431f-999a-4213-a745-9d8b03aa028e-kube-api-access-r8stp\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.381806 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8daa431f-999a-4213-a745-9d8b03aa028e-config-volume\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.395153 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8daa431f-999a-4213-a745-9d8b03aa028e-secret-volume\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.398050 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8stp\" (UniqueName: \"kubernetes.io/projected/8daa431f-999a-4213-a745-9d8b03aa028e-kube-api-access-r8stp\") pod \"collect-profiles-29413830-kqvz4\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.442444 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 04 06:30:00 crc kubenswrapper[4685]: I1204 06:30:00.495990 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.631525 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.631774 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rf9kv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-cpkd7_openstack(6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.632893 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-cpkd7" podUID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.646216 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.646551 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n599h554h56fh545h5ch569h95h645h9h54bh67h65h7bh659h5f8h66dh646h5b9h9fhb9h57fh5cdh584h68h97h68fh5b8h677h698hd9h75h67fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-464vx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6fb76c6c8c-j6b94_openstack(6d9a855f-3727-41c7-85a9-d404efbd4d83): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.649316 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6fb76c6c8c-j6b94" podUID="6d9a855f-3727-41c7-85a9-d404efbd4d83" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.653924 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.654049 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d8h59fh585hcbhdh7fh64bh9dh65fh5fbh5f7h55h85hddh56fh66h574h5fbh56dh657hbdh598h75h558h5d9h5ddhcfh59bh67hbfh596h65q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jgkrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5588ffff99-v5xfx_openstack(5fa411a1-1dc0-4140-b7f7-8684304402a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:30:01 crc kubenswrapper[4685]: E1204 06:30:01.658634 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5588ffff99-v5xfx" podUID="5fa411a1-1dc0-4140-b7f7-8684304402a5" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.739016 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.909790 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-combined-ca-bundle\") pod \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.909882 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-fernet-keys\") pod \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.910017 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-config-data\") pod \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.910060 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp4n8\" (UniqueName: \"kubernetes.io/projected/25dfcb44-a95f-466e-b9fe-f8080cea09e8-kube-api-access-lp4n8\") pod \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.910087 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-credential-keys\") pod \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.910140 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-scripts\") pod \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\" (UID: \"25dfcb44-a95f-466e-b9fe-f8080cea09e8\") " Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.916375 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "25dfcb44-a95f-466e-b9fe-f8080cea09e8" (UID: "25dfcb44-a95f-466e-b9fe-f8080cea09e8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.916472 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25dfcb44-a95f-466e-b9fe-f8080cea09e8-kube-api-access-lp4n8" (OuterVolumeSpecName: "kube-api-access-lp4n8") pod "25dfcb44-a95f-466e-b9fe-f8080cea09e8" (UID: "25dfcb44-a95f-466e-b9fe-f8080cea09e8"). InnerVolumeSpecName "kube-api-access-lp4n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.916503 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-scripts" (OuterVolumeSpecName: "scripts") pod "25dfcb44-a95f-466e-b9fe-f8080cea09e8" (UID: "25dfcb44-a95f-466e-b9fe-f8080cea09e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.918187 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "25dfcb44-a95f-466e-b9fe-f8080cea09e8" (UID: "25dfcb44-a95f-466e-b9fe-f8080cea09e8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.936330 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-config-data" (OuterVolumeSpecName: "config-data") pod "25dfcb44-a95f-466e-b9fe-f8080cea09e8" (UID: "25dfcb44-a95f-466e-b9fe-f8080cea09e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:01 crc kubenswrapper[4685]: I1204 06:30:01.936643 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25dfcb44-a95f-466e-b9fe-f8080cea09e8" (UID: "25dfcb44-a95f-466e-b9fe-f8080cea09e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.012337 4685 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.012374 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.012388 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp4n8\" (UniqueName: \"kubernetes.io/projected/25dfcb44-a95f-466e-b9fe-f8080cea09e8-kube-api-access-lp4n8\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.012404 4685 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.012462 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.012473 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25dfcb44-a95f-466e-b9fe-f8080cea09e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.579006 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkpwl" event={"ID":"25dfcb44-a95f-466e-b9fe-f8080cea09e8","Type":"ContainerDied","Data":"f27ec6355626100c05d9ec57e354ceddd4ab4fb43c70e1d3797a2d9103ea75a4"} Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.579058 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkpwl" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.579159 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f27ec6355626100c05d9ec57e354ceddd4ab4fb43c70e1d3797a2d9103ea75a4" Dec 04 06:30:02 crc kubenswrapper[4685]: E1204 06:30:02.582803 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-cpkd7" podUID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.832197 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nkpwl"] Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.839667 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nkpwl"] Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.918810 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bljbv"] Dec 04 06:30:02 crc kubenswrapper[4685]: E1204 06:30:02.919316 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25dfcb44-a95f-466e-b9fe-f8080cea09e8" containerName="keystone-bootstrap" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.919331 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="25dfcb44-a95f-466e-b9fe-f8080cea09e8" containerName="keystone-bootstrap" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.919553 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="25dfcb44-a95f-466e-b9fe-f8080cea09e8" containerName="keystone-bootstrap" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.920203 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.922919 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.922965 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.923016 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.923063 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c6l2p" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.923526 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 06:30:02 crc kubenswrapper[4685]: I1204 06:30:02.939482 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bljbv"] Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.032149 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-credential-keys\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.032214 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzvc\" (UniqueName: \"kubernetes.io/projected/fd072a3b-1456-43ad-97a7-7326a66a6f9d-kube-api-access-zmzvc\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.032475 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-config-data\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.032595 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-combined-ca-bundle\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.032632 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-scripts\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.032657 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-fernet-keys\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.134098 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-credential-keys\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.134162 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzvc\" (UniqueName: \"kubernetes.io/projected/fd072a3b-1456-43ad-97a7-7326a66a6f9d-kube-api-access-zmzvc\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.134207 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-config-data\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.134264 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-combined-ca-bundle\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.134283 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-scripts\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.134299 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-fernet-keys\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.136605 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.136605 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.137139 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.140401 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-combined-ca-bundle\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.149439 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25dfcb44-a95f-466e-b9fe-f8080cea09e8" path="/var/lib/kubelet/pods/25dfcb44-a95f-466e-b9fe-f8080cea09e8/volumes" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.154347 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-scripts\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.155562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-credential-keys\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.155559 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-config-data\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.155677 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-fernet-keys\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.166335 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzvc\" (UniqueName: \"kubernetes.io/projected/fd072a3b-1456-43ad-97a7-7326a66a6f9d-kube-api-access-zmzvc\") pod \"keystone-bootstrap-bljbv\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.243347 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c6l2p" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.250114 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.588802 4685 generic.go:334] "Generic (PLEG): container finished" podID="4ab9f58d-e6bb-445d-a748-0059a8981f49" containerID="9472458b1f7b017c0ddb1d80493683f3e4e415e4cdc6387da6cd9b1b21bd7729" exitCode=0 Dec 04 06:30:03 crc kubenswrapper[4685]: I1204 06:30:03.588846 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcqrm" event={"ID":"4ab9f58d-e6bb-445d-a748-0059a8981f49","Type":"ContainerDied","Data":"9472458b1f7b017c0ddb1d80493683f3e4e415e4cdc6387da6cd9b1b21bd7729"} Dec 04 06:30:05 crc kubenswrapper[4685]: I1204 06:30:05.442774 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 04 06:30:05 crc kubenswrapper[4685]: I1204 06:30:05.442969 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.121602 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.121917 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.121991 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.122634 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bca40440ad9d82bdf8b013f3ae5e8f16199c54c667e88ee4d71345b125617d55"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.122711 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://bca40440ad9d82bdf8b013f3ae5e8f16199c54c667e88ee4d71345b125617d55" gracePeriod=600 Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.626815 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="bca40440ad9d82bdf8b013f3ae5e8f16199c54c667e88ee4d71345b125617d55" exitCode=0 Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.626887 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"bca40440ad9d82bdf8b013f3ae5e8f16199c54c667e88ee4d71345b125617d55"} Dec 04 06:30:06 crc kubenswrapper[4685]: I1204 06:30:06.626972 4685 scope.go:117] "RemoveContainer" containerID="1bbd856de92fdeb4119a37d3bc3ff75b3516aafca468b858671a77ab07bd5c11" Dec 04 06:30:09 crc kubenswrapper[4685]: I1204 06:30:09.962899 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:09 crc kubenswrapper[4685]: I1204 06:30:09.970961 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-scripts\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082347 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-logs\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082372 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-scripts\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082522 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-httpd-run\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082566 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-combined-ca-bundle\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082595 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssnzm\" (UniqueName: \"kubernetes.io/projected/98de6ea0-e984-4319-a585-d70e40f4291f-kube-api-access-ssnzm\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082637 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-combined-ca-bundle\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082654 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-config-data\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082684 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-config-data\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082796 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdxjl\" (UniqueName: \"kubernetes.io/projected/f908631b-9f25-474e-9dff-815a7dc11aae-kube-api-access-pdxjl\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082839 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-logs\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082881 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082911 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082946 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-httpd-run\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.082982 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-public-tls-certs\") pod \"f908631b-9f25-474e-9dff-815a7dc11aae\" (UID: \"f908631b-9f25-474e-9dff-815a7dc11aae\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.083018 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-internal-tls-certs\") pod \"98de6ea0-e984-4319-a585-d70e40f4291f\" (UID: \"98de6ea0-e984-4319-a585-d70e40f4291f\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.088709 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-logs" (OuterVolumeSpecName: "logs") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.089041 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-logs" (OuterVolumeSpecName: "logs") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.089306 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.091021 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.092373 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-scripts" (OuterVolumeSpecName: "scripts") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.103624 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-scripts" (OuterVolumeSpecName: "scripts") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.104060 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98de6ea0-e984-4319-a585-d70e40f4291f-kube-api-access-ssnzm" (OuterVolumeSpecName: "kube-api-access-ssnzm") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "kube-api-access-ssnzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.105631 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.105659 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f908631b-9f25-474e-9dff-815a7dc11aae-kube-api-access-pdxjl" (OuterVolumeSpecName: "kube-api-access-pdxjl") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "kube-api-access-pdxjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.133126 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.155298 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184663 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184694 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184703 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184711 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184720 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssnzm\" (UniqueName: \"kubernetes.io/projected/98de6ea0-e984-4319-a585-d70e40f4291f-kube-api-access-ssnzm\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184732 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184740 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdxjl\" (UniqueName: \"kubernetes.io/projected/f908631b-9f25-474e-9dff-815a7dc11aae-kube-api-access-pdxjl\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184749 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f908631b-9f25-474e-9dff-815a7dc11aae-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184775 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184787 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.184796 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98de6ea0-e984-4319-a585-d70e40f4291f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.191532 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.192208 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.195954 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-config-data" (OuterVolumeSpecName: "config-data") pod "98de6ea0-e984-4319-a585-d70e40f4291f" (UID: "98de6ea0-e984-4319-a585-d70e40f4291f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.203068 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.224553 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-config-data" (OuterVolumeSpecName: "config-data") pod "f908631b-9f25-474e-9dff-815a7dc11aae" (UID: "f908631b-9f25-474e-9dff-815a7dc11aae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.251634 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.291161 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.291389 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.291465 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.291517 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.291565 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f908631b-9f25-474e-9dff-815a7dc11aae-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.291611 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98de6ea0-e984-4319-a585-d70e40f4291f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.325336 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.392789 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.672687 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.672688 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f908631b-9f25-474e-9dff-815a7dc11aae","Type":"ContainerDied","Data":"ba797b8fe42614311931fca52ba0a5ac490d81901c8b947bb83fe6159edfe41a"} Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.677640 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98de6ea0-e984-4319-a585-d70e40f4291f","Type":"ContainerDied","Data":"b97bd201ac1e78892db145b586d5065803b23400ff62e05ad36bd45865c83c8f"} Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.677789 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.705840 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.706017 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-48pl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-5498d_openstack(6f6bce91-0f30-4543-ac33-b34877206b05): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.707643 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-5498d" podUID="6f6bce91-0f30-4543-ac33-b34877206b05" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.718446 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.735401 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.748464 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.755286 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.761369 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.769560 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.786841 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.838096 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.839538 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-log" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.839575 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-log" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.839864 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-httpd" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.839899 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-httpd" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.839924 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab9f58d-e6bb-445d-a748-0059a8981f49" containerName="neutron-db-sync" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.839930 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab9f58d-e6bb-445d-a748-0059a8981f49" containerName="neutron-db-sync" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.839947 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-log" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.839953 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-log" Dec 04 06:30:10 crc kubenswrapper[4685]: E1204 06:30:10.839966 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-httpd" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.839971 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-httpd" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.840206 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-httpd" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.840221 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-log" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.840235 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" containerName="glance-log" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.840256 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" containerName="glance-httpd" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.840268 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab9f58d-e6bb-445d-a748-0059a8981f49" containerName="neutron-db-sync" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.843737 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.854188 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.854441 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-66ml8" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.854580 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.854709 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.870167 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.871871 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.880944 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.886887 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.887063 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.904884 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5fa411a1-1dc0-4140-b7f7-8684304402a5-horizon-secret-key\") pod \"5fa411a1-1dc0-4140-b7f7-8684304402a5\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.904963 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d9a855f-3727-41c7-85a9-d404efbd4d83-horizon-secret-key\") pod \"6d9a855f-3727-41c7-85a9-d404efbd4d83\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905018 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qxl4\" (UniqueName: \"kubernetes.io/projected/4ab9f58d-e6bb-445d-a748-0059a8981f49-kube-api-access-8qxl4\") pod \"4ab9f58d-e6bb-445d-a748-0059a8981f49\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905053 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-config-data\") pod \"6d9a855f-3727-41c7-85a9-d404efbd4d83\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905087 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d9a855f-3727-41c7-85a9-d404efbd4d83-logs\") pod \"6d9a855f-3727-41c7-85a9-d404efbd4d83\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905187 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-scripts\") pod \"5fa411a1-1dc0-4140-b7f7-8684304402a5\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905203 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-config\") pod \"4ab9f58d-e6bb-445d-a748-0059a8981f49\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905233 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fa411a1-1dc0-4140-b7f7-8684304402a5-logs\") pod \"5fa411a1-1dc0-4140-b7f7-8684304402a5\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905252 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-scripts\") pod \"6d9a855f-3727-41c7-85a9-d404efbd4d83\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905274 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-464vx\" (UniqueName: \"kubernetes.io/projected/6d9a855f-3727-41c7-85a9-d404efbd4d83-kube-api-access-464vx\") pod \"6d9a855f-3727-41c7-85a9-d404efbd4d83\" (UID: \"6d9a855f-3727-41c7-85a9-d404efbd4d83\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905318 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgkrf\" (UniqueName: \"kubernetes.io/projected/5fa411a1-1dc0-4140-b7f7-8684304402a5-kube-api-access-jgkrf\") pod \"5fa411a1-1dc0-4140-b7f7-8684304402a5\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905352 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-config-data\") pod \"5fa411a1-1dc0-4140-b7f7-8684304402a5\" (UID: \"5fa411a1-1dc0-4140-b7f7-8684304402a5\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.905371 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-combined-ca-bundle\") pod \"4ab9f58d-e6bb-445d-a748-0059a8981f49\" (UID: \"4ab9f58d-e6bb-445d-a748-0059a8981f49\") " Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.906243 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.906986 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-scripts" (OuterVolumeSpecName: "scripts") pod "5fa411a1-1dc0-4140-b7f7-8684304402a5" (UID: "5fa411a1-1dc0-4140-b7f7-8684304402a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.907029 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-scripts" (OuterVolumeSpecName: "scripts") pod "6d9a855f-3727-41c7-85a9-d404efbd4d83" (UID: "6d9a855f-3727-41c7-85a9-d404efbd4d83"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.909735 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fa411a1-1dc0-4140-b7f7-8684304402a5-logs" (OuterVolumeSpecName: "logs") pod "5fa411a1-1dc0-4140-b7f7-8684304402a5" (UID: "5fa411a1-1dc0-4140-b7f7-8684304402a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.910035 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-config-data" (OuterVolumeSpecName: "config-data") pod "6d9a855f-3727-41c7-85a9-d404efbd4d83" (UID: "6d9a855f-3727-41c7-85a9-d404efbd4d83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.910798 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d9a855f-3727-41c7-85a9-d404efbd4d83-logs" (OuterVolumeSpecName: "logs") pod "6d9a855f-3727-41c7-85a9-d404efbd4d83" (UID: "6d9a855f-3727-41c7-85a9-d404efbd4d83"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.910955 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa411a1-1dc0-4140-b7f7-8684304402a5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5fa411a1-1dc0-4140-b7f7-8684304402a5" (UID: "5fa411a1-1dc0-4140-b7f7-8684304402a5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.911825 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-config-data" (OuterVolumeSpecName: "config-data") pod "5fa411a1-1dc0-4140-b7f7-8684304402a5" (UID: "5fa411a1-1dc0-4140-b7f7-8684304402a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.917216 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9a855f-3727-41c7-85a9-d404efbd4d83-kube-api-access-464vx" (OuterVolumeSpecName: "kube-api-access-464vx") pod "6d9a855f-3727-41c7-85a9-d404efbd4d83" (UID: "6d9a855f-3727-41c7-85a9-d404efbd4d83"). InnerVolumeSpecName "kube-api-access-464vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.920355 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9a855f-3727-41c7-85a9-d404efbd4d83-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6d9a855f-3727-41c7-85a9-d404efbd4d83" (UID: "6d9a855f-3727-41c7-85a9-d404efbd4d83"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.921725 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab9f58d-e6bb-445d-a748-0059a8981f49-kube-api-access-8qxl4" (OuterVolumeSpecName: "kube-api-access-8qxl4") pod "4ab9f58d-e6bb-445d-a748-0059a8981f49" (UID: "4ab9f58d-e6bb-445d-a748-0059a8981f49"). InnerVolumeSpecName "kube-api-access-8qxl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.928098 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa411a1-1dc0-4140-b7f7-8684304402a5-kube-api-access-jgkrf" (OuterVolumeSpecName: "kube-api-access-jgkrf") pod "5fa411a1-1dc0-4140-b7f7-8684304402a5" (UID: "5fa411a1-1dc0-4140-b7f7-8684304402a5"). InnerVolumeSpecName "kube-api-access-jgkrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.945883 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-config" (OuterVolumeSpecName: "config") pod "4ab9f58d-e6bb-445d-a748-0059a8981f49" (UID: "4ab9f58d-e6bb-445d-a748-0059a8981f49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:10 crc kubenswrapper[4685]: I1204 06:30:10.968610 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ab9f58d-e6bb-445d-a748-0059a8981f49" (UID: "4ab9f58d-e6bb-445d-a748-0059a8981f49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007359 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007422 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007445 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7z7d\" (UniqueName: \"kubernetes.io/projected/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-kube-api-access-c7z7d\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007472 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007492 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007531 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007560 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007610 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgpfn\" (UniqueName: \"kubernetes.io/projected/5d114929-218b-465a-9116-ed8144ed0636-kube-api-access-bgpfn\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007643 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007664 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-config-data\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007684 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-logs\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007700 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-scripts\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007723 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007742 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.007957 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5fa411a1-1dc0-4140-b7f7-8684304402a5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008004 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d9a855f-3727-41c7-85a9-d404efbd4d83-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008016 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qxl4\" (UniqueName: \"kubernetes.io/projected/4ab9f58d-e6bb-445d-a748-0059a8981f49-kube-api-access-8qxl4\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008027 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008039 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d9a855f-3727-41c7-85a9-d404efbd4d83-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008051 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008060 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008071 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fa411a1-1dc0-4140-b7f7-8684304402a5-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008121 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d9a855f-3727-41c7-85a9-d404efbd4d83-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008133 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-464vx\" (UniqueName: \"kubernetes.io/projected/6d9a855f-3727-41c7-85a9-d404efbd4d83-kube-api-access-464vx\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008144 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgkrf\" (UniqueName: \"kubernetes.io/projected/5fa411a1-1dc0-4140-b7f7-8684304402a5-kube-api-access-jgkrf\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008154 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fa411a1-1dc0-4140-b7f7-8684304402a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.008163 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab9f58d-e6bb-445d-a748-0059a8981f49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.111210 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.111384 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.111476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7z7d\" (UniqueName: \"kubernetes.io/projected/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-kube-api-access-c7z7d\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.111668 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114048 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114123 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114149 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114209 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114297 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114306 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.114328 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.115218 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.115898 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116164 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116357 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgpfn\" (UniqueName: \"kubernetes.io/projected/5d114929-218b-465a-9116-ed8144ed0636-kube-api-access-bgpfn\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116499 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116539 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-config-data\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116582 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-logs\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-scripts\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116655 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.116692 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.118987 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.119253 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-logs\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.123156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.125231 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.126067 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.126486 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.127219 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-config-data\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.128986 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.129866 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-scripts\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.137819 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7z7d\" (UniqueName: \"kubernetes.io/projected/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-kube-api-access-c7z7d\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.140666 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgpfn\" (UniqueName: \"kubernetes.io/projected/5d114929-218b-465a-9116-ed8144ed0636-kube-api-access-bgpfn\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.151944 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.153386 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98de6ea0-e984-4319-a585-d70e40f4291f" path="/var/lib/kubelet/pods/98de6ea0-e984-4319-a585-d70e40f4291f/volumes" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.154332 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f908631b-9f25-474e-9dff-815a7dc11aae" path="/var/lib/kubelet/pods/f908631b-9f25-474e-9dff-815a7dc11aae/volumes" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.172265 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.343112 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.351839 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.689985 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5588ffff99-v5xfx" event={"ID":"5fa411a1-1dc0-4140-b7f7-8684304402a5","Type":"ContainerDied","Data":"06bda29bf5c9e409e287ad60f9e365248b522bfdd44fa040a8b10b778986f22e"} Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.690029 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5588ffff99-v5xfx" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.692425 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcqrm" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.692401 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcqrm" event={"ID":"4ab9f58d-e6bb-445d-a748-0059a8981f49","Type":"ContainerDied","Data":"066a7793eec3e8427e39f0770c8ff919fbd93b10dd22c3f4c4a3320545ba35df"} Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.692654 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="066a7793eec3e8427e39f0770c8ff919fbd93b10dd22c3f4c4a3320545ba35df" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.708281 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb76c6c8c-j6b94" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.708760 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb76c6c8c-j6b94" event={"ID":"6d9a855f-3727-41c7-85a9-d404efbd4d83","Type":"ContainerDied","Data":"92a6ab637515273894234fa0a3bd94d01d819ece1085e1c30bdd6805eb6e74b3"} Dec 04 06:30:11 crc kubenswrapper[4685]: E1204 06:30:11.710293 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-5498d" podUID="6f6bce91-0f30-4543-ac33-b34877206b05" Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.749547 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5588ffff99-v5xfx"] Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.770922 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5588ffff99-v5xfx"] Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.801474 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fb76c6c8c-j6b94"] Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.812081 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6fb76c6c8c-j6b94"] Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.981349 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-tvq7r"] Dec 04 06:30:11 crc kubenswrapper[4685]: I1204 06:30:11.983008 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.000866 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-tvq7r"] Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.036278 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.036354 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.036414 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zz5b\" (UniqueName: \"kubernetes.io/projected/c52e4c01-ae05-4a0b-b245-04e570525017-kube-api-access-2zz5b\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.036434 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-svc\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.036452 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.036472 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-config\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.106786 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6985cd56b8-9gnlv"] Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.111992 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.121881 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.121895 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-p9tjg" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.121994 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.122554 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.139213 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.139301 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.139370 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zz5b\" (UniqueName: \"kubernetes.io/projected/c52e4c01-ae05-4a0b-b245-04e570525017-kube-api-access-2zz5b\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.139389 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-svc\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.139435 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.139458 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-config\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.144568 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.145608 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.145803 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-svc\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.146249 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.146668 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-config\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.149888 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6985cd56b8-9gnlv"] Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.206187 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zz5b\" (UniqueName: \"kubernetes.io/projected/c52e4c01-ae05-4a0b-b245-04e570525017-kube-api-access-2zz5b\") pod \"dnsmasq-dns-6b7b667979-tvq7r\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.244653 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-httpd-config\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.244874 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-ovndb-tls-certs\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.245119 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-config\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.245289 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzzp6\" (UniqueName: \"kubernetes.io/projected/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-kube-api-access-wzzp6\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.245441 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-combined-ca-bundle\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.314238 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.350149 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-combined-ca-bundle\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.350270 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-httpd-config\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.350461 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-ovndb-tls-certs\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.350557 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-config\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.350647 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzzp6\" (UniqueName: \"kubernetes.io/projected/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-kube-api-access-wzzp6\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.353948 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-httpd-config\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.354367 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-config\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.356162 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-ovndb-tls-certs\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.368081 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzzp6\" (UniqueName: \"kubernetes.io/projected/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-kube-api-access-wzzp6\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.372770 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-combined-ca-bundle\") pod \"neutron-6985cd56b8-9gnlv\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:12 crc kubenswrapper[4685]: I1204 06:30:12.463783 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:13 crc kubenswrapper[4685]: I1204 06:30:13.135090 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa411a1-1dc0-4140-b7f7-8684304402a5" path="/var/lib/kubelet/pods/5fa411a1-1dc0-4140-b7f7-8684304402a5/volumes" Dec 04 06:30:13 crc kubenswrapper[4685]: I1204 06:30:13.135820 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9a855f-3727-41c7-85a9-d404efbd4d83" path="/var/lib/kubelet/pods/6d9a855f-3727-41c7-85a9-d404efbd4d83/volumes" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.610824 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-964564b4f-vkm6n"] Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.612601 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: E1204 06:30:14.632700 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 04 06:30:14 crc kubenswrapper[4685]: E1204 06:30:14.632928 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-psn2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-59cdh_openstack(964316c7-6da5-4725-8140-6d249ee01a68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.633502 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.633761 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 04 06:30:14 crc kubenswrapper[4685]: E1204 06:30:14.642171 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-59cdh" podUID="964316c7-6da5-4725-8140-6d249ee01a68" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.643937 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-964564b4f-vkm6n"] Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.700797 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz4bv\" (UniqueName: \"kubernetes.io/projected/0452cb7f-8545-42b9-9456-b29aeaf9c391-kube-api-access-hz4bv\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.701106 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-config\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.701278 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-httpd-config\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.701313 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-internal-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.701347 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-public-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.701376 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-combined-ca-bundle\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.701416 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-ovndb-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.745942 4685 scope.go:117] "RemoveContainer" containerID="1f05d49f2d3688095750e25d589e7541d9586496e29a620e646130ff778717c0" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.775993 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" event={"ID":"991d5443-6323-48ba-9a55-832b4d80820b","Type":"ContainerDied","Data":"88b62b262eca01b36fcbd31c9e70952a2ce38874d6a7f906bc1109e51f732f10"} Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.776039 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b62b262eca01b36fcbd31c9e70952a2ce38874d6a7f906bc1109e51f732f10" Dec 04 06:30:14 crc kubenswrapper[4685]: E1204 06:30:14.777916 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-59cdh" podUID="964316c7-6da5-4725-8140-6d249ee01a68" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803103 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-httpd-config\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803172 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-internal-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803218 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-public-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803267 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-combined-ca-bundle\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803295 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-ovndb-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803337 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz4bv\" (UniqueName: \"kubernetes.io/projected/0452cb7f-8545-42b9-9456-b29aeaf9c391-kube-api-access-hz4bv\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.803390 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-config\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.814079 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-combined-ca-bundle\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.815634 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-httpd-config\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.817744 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-public-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.832058 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-config\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.836912 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.840996 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-ovndb-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.863097 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0452cb7f-8545-42b9-9456-b29aeaf9c391-internal-tls-certs\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.869169 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz4bv\" (UniqueName: \"kubernetes.io/projected/0452cb7f-8545-42b9-9456-b29aeaf9c391-kube-api-access-hz4bv\") pod \"neutron-964564b4f-vkm6n\" (UID: \"0452cb7f-8545-42b9-9456-b29aeaf9c391\") " pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.904964 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-svc\") pod \"991d5443-6323-48ba-9a55-832b4d80820b\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.905095 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-swift-storage-0\") pod \"991d5443-6323-48ba-9a55-832b4d80820b\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.905160 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlwll\" (UniqueName: \"kubernetes.io/projected/991d5443-6323-48ba-9a55-832b4d80820b-kube-api-access-nlwll\") pod \"991d5443-6323-48ba-9a55-832b4d80820b\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.905202 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-nb\") pod \"991d5443-6323-48ba-9a55-832b4d80820b\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.905242 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-sb\") pod \"991d5443-6323-48ba-9a55-832b4d80820b\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.905272 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-config\") pod \"991d5443-6323-48ba-9a55-832b4d80820b\" (UID: \"991d5443-6323-48ba-9a55-832b4d80820b\") " Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.932775 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/991d5443-6323-48ba-9a55-832b4d80820b-kube-api-access-nlwll" (OuterVolumeSpecName: "kube-api-access-nlwll") pod "991d5443-6323-48ba-9a55-832b4d80820b" (UID: "991d5443-6323-48ba-9a55-832b4d80820b"). InnerVolumeSpecName "kube-api-access-nlwll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:14 crc kubenswrapper[4685]: I1204 06:30:14.955845 4685 scope.go:117] "RemoveContainer" containerID="6e0ad47fe1e6118f1a881b2cc0b01072646fbbc4f857be28eff12ec66a93a635" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:14.999817 4685 scope.go:117] "RemoveContainer" containerID="e5de6316c344d13be595a2691d616ff6e53470ddf3bbdac90998471d89c80a50" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.007937 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlwll\" (UniqueName: \"kubernetes.io/projected/991d5443-6323-48ba-9a55-832b4d80820b-kube-api-access-nlwll\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.013461 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "991d5443-6323-48ba-9a55-832b4d80820b" (UID: "991d5443-6323-48ba-9a55-832b4d80820b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.078731 4685 scope.go:117] "RemoveContainer" containerID="c912f3f748574799ef519291dc29aa36647a8cee1888485f19e13ccdead9b27d" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.094930 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bdfd7d6f6-8slln"] Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.110089 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.126938 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "991d5443-6323-48ba-9a55-832b4d80820b" (UID: "991d5443-6323-48ba-9a55-832b4d80820b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.138762 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "991d5443-6323-48ba-9a55-832b4d80820b" (UID: "991d5443-6323-48ba-9a55-832b4d80820b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.140035 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "991d5443-6323-48ba-9a55-832b4d80820b" (UID: "991d5443-6323-48ba-9a55-832b4d80820b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.155646 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.165064 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-config" (OuterVolumeSpecName: "config") pod "991d5443-6323-48ba-9a55-832b4d80820b" (UID: "991d5443-6323-48ba-9a55-832b4d80820b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.211785 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.212236 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.212294 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.212346 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991d5443-6323-48ba-9a55-832b4d80820b-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.441697 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.547060 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c9548d9dd-h5vqf"] Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.818779 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bdfd7d6f6-8slln" event={"ID":"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8","Type":"ContainerStarted","Data":"252f6828706c15e58fcb7a3b491ecee0a37c85fd571fbefe50fcd5492a05e612"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.819083 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bdfd7d6f6-8slln" event={"ID":"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8","Type":"ContainerStarted","Data":"9e9c684185854ea65eeabdfa138d41f84665bb035420a994e2d88229d118c302"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.821251 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddd9d7c47-dwpfz" event={"ID":"b56721cb-e8ce-4106-920b-0b735ef29a38","Type":"ContainerStarted","Data":"5fba8a68b2f3c8900cd7c2d81f68ffadb6367ed2359e9e2be619f08f13c5ef30"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.821275 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddd9d7c47-dwpfz" event={"ID":"b56721cb-e8ce-4106-920b-0b735ef29a38","Type":"ContainerStarted","Data":"c78490ec1b2576e8d2d66ed40080d3ef47ad819e6a75fa496653827bd45b1d9c"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.821424 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6ddd9d7c47-dwpfz" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon-log" containerID="cri-o://c78490ec1b2576e8d2d66ed40080d3ef47ad819e6a75fa496653827bd45b1d9c" gracePeriod=30 Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.821886 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6ddd9d7c47-dwpfz" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon" containerID="cri-o://5fba8a68b2f3c8900cd7c2d81f68ffadb6367ed2359e9e2be619f08f13c5ef30" gracePeriod=30 Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.829303 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c9548d9dd-h5vqf" event={"ID":"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236","Type":"ContainerStarted","Data":"7beff68097eb7d4f2ba36744fa546a10a9b298787c6d99f6e67f58c7977468e2"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.842559 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6ddd9d7c47-dwpfz" podStartSLOduration=7.094173028 podStartE2EDuration="30.842537229s" podCreationTimestamp="2025-12-04 06:29:45 +0000 UTC" firstStartedPulling="2025-12-04 06:29:46.945581704 +0000 UTC m=+1124.133812479" lastFinishedPulling="2025-12-04 06:30:10.693945905 +0000 UTC m=+1147.882176680" observedRunningTime="2025-12-04 06:30:15.839944847 +0000 UTC m=+1153.028175622" watchObservedRunningTime="2025-12-04 06:30:15.842537229 +0000 UTC m=+1153.030768004" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.847293 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"6fc1b67a51a8fb7e262f18c32ce16b7ebc33919ea2c7327f8e192aec0ec81fc6"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.860630 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerStarted","Data":"56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3"} Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.867737 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9wqf7" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.879355 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bljbv"] Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.909818 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.927708 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4"] Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.938237 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9wqf7"] Dec 04 06:30:15 crc kubenswrapper[4685]: I1204 06:30:15.944511 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9wqf7"] Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.120738 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.160551 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.256498 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-tvq7r"] Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.332013 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-964564b4f-vkm6n"] Dec 04 06:30:16 crc kubenswrapper[4685]: W1204 06:30:16.342674 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0452cb7f_8545_42b9_9456_b29aeaf9c391.slice/crio-28825cbbace88de732a6a8d18201ec641012cdcf64c8ebab52d6ebf8b6944154 WatchSource:0}: Error finding container 28825cbbace88de732a6a8d18201ec641012cdcf64c8ebab52d6ebf8b6944154: Status 404 returned error can't find the container with id 28825cbbace88de732a6a8d18201ec641012cdcf64c8ebab52d6ebf8b6944154 Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.908282 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-964564b4f-vkm6n" event={"ID":"0452cb7f-8545-42b9-9456-b29aeaf9c391","Type":"ContainerStarted","Data":"e5539313598960d95a254fc07b344876107295acc26e2c579e383d2218f48f04"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.918975 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-964564b4f-vkm6n" event={"ID":"0452cb7f-8545-42b9-9456-b29aeaf9c391","Type":"ContainerStarted","Data":"28825cbbace88de732a6a8d18201ec641012cdcf64c8ebab52d6ebf8b6944154"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.919607 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cpkd7" event={"ID":"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3","Type":"ContainerStarted","Data":"dfe3ec87dcc103e750d55dae54a2e0d01f746764027360ddef79b6fa37fbfa75"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.922518 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5d114929-218b-465a-9116-ed8144ed0636","Type":"ContainerStarted","Data":"131407d07c03df3460693da013ab6d63e82d2b24d4c73956e55aacee1f753362"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.934613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c9548d9dd-h5vqf" event={"ID":"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236","Type":"ContainerStarted","Data":"abe49f70548322230db3ef0f78f81934821f85393fb61bb10f09d332bc37865a"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.934945 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c9548d9dd-h5vqf" event={"ID":"0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236","Type":"ContainerStarted","Data":"e5d70b787d7cdc038f071038ba3c0d95e92b683e4ff65f3febbe1d99d01f8062"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.950182 4685 generic.go:334] "Generic (PLEG): container finished" podID="8daa431f-999a-4213-a745-9d8b03aa028e" containerID="55c52dcec029a86895ba594f836748f637c993c60370e69352ae95b6419ec4ef" exitCode=0 Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.950263 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" event={"ID":"8daa431f-999a-4213-a745-9d8b03aa028e","Type":"ContainerDied","Data":"55c52dcec029a86895ba594f836748f637c993c60370e69352ae95b6419ec4ef"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.950289 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" event={"ID":"8daa431f-999a-4213-a745-9d8b03aa028e","Type":"ContainerStarted","Data":"38a4c2a6d92a441725360a97988e3ace57b332715413faa7ce883d0c1336a522"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.956868 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cpkd7" podStartSLOduration=3.424549431 podStartE2EDuration="33.956842299s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="2025-12-04 06:29:45.208625563 +0000 UTC m=+1122.396856338" lastFinishedPulling="2025-12-04 06:30:15.740918431 +0000 UTC m=+1152.929149206" observedRunningTime="2025-12-04 06:30:16.942061104 +0000 UTC m=+1154.130291879" watchObservedRunningTime="2025-12-04 06:30:16.956842299 +0000 UTC m=+1154.145073074" Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.960273 4685 generic.go:334] "Generic (PLEG): container finished" podID="c52e4c01-ae05-4a0b-b245-04e570525017" containerID="6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af" exitCode=0 Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.960445 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" event={"ID":"c52e4c01-ae05-4a0b-b245-04e570525017","Type":"ContainerDied","Data":"6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.960548 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" event={"ID":"c52e4c01-ae05-4a0b-b245-04e570525017","Type":"ContainerStarted","Data":"74144e9ce8665bade7060dca11978b27a2bc5c59b9bd259caa8f26ee2f42bd52"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.977002 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bljbv" event={"ID":"fd072a3b-1456-43ad-97a7-7326a66a6f9d","Type":"ContainerStarted","Data":"eddd8dea3e1ba7f7c2f7b5188170acda9d987dc8ab31a43c1061e616a86e8b30"} Dec 04 06:30:16 crc kubenswrapper[4685]: I1204 06:30:16.977052 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bljbv" event={"ID":"fd072a3b-1456-43ad-97a7-7326a66a6f9d","Type":"ContainerStarted","Data":"bada0e63685a17346e53af214b5c758197e63770fd8d84dc7c499c0e78bf4d10"} Dec 04 06:30:17 crc kubenswrapper[4685]: I1204 06:30:17.020598 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c9548d9dd-h5vqf" podStartSLOduration=25.020567054 podStartE2EDuration="25.020567054s" podCreationTimestamp="2025-12-04 06:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:16.96798406 +0000 UTC m=+1154.156214855" watchObservedRunningTime="2025-12-04 06:30:17.020567054 +0000 UTC m=+1154.208797819" Dec 04 06:30:17 crc kubenswrapper[4685]: I1204 06:30:17.049355 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bdfd7d6f6-8slln" event={"ID":"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8","Type":"ContainerStarted","Data":"7195b2be6943bd0983daf635bd5ec023782817786ac253eb69ce2b406968b974"} Dec 04 06:30:17 crc kubenswrapper[4685]: I1204 06:30:17.065491 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:30:17 crc kubenswrapper[4685]: I1204 06:30:17.102188 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bljbv" podStartSLOduration=15.102165571 podStartE2EDuration="15.102165571s" podCreationTimestamp="2025-12-04 06:30:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:17.060950285 +0000 UTC m=+1154.249181060" watchObservedRunningTime="2025-12-04 06:30:17.102165571 +0000 UTC m=+1154.290396346" Dec 04 06:30:17 crc kubenswrapper[4685]: I1204 06:30:17.116796 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6bdfd7d6f6-8slln" podStartSLOduration=25.116776501 podStartE2EDuration="25.116776501s" podCreationTimestamp="2025-12-04 06:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:17.088065278 +0000 UTC m=+1154.276296053" watchObservedRunningTime="2025-12-04 06:30:17.116776501 +0000 UTC m=+1154.305007276" Dec 04 06:30:17 crc kubenswrapper[4685]: I1204 06:30:17.168011 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="991d5443-6323-48ba-9a55-832b4d80820b" path="/var/lib/kubelet/pods/991d5443-6323-48ba-9a55-832b4d80820b/volumes" Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.028467 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6985cd56b8-9gnlv"] Dec 04 06:30:18 crc kubenswrapper[4685]: W1204 06:30:18.114819 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc657cb75_b17e_41fb_8192_2a6cf3ccbe85.slice/crio-e2d727cf7329731ea689cc3001d03f5971833f132743fc5beb636e3aca5e23aa WatchSource:0}: Error finding container e2d727cf7329731ea689cc3001d03f5971833f132743fc5beb636e3aca5e23aa: Status 404 returned error can't find the container with id e2d727cf7329731ea689cc3001d03f5971833f132743fc5beb636e3aca5e23aa Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.121537 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34","Type":"ContainerStarted","Data":"bda258831aeb73f79d09ceafebd3635a2181f184109f19edc1c368eb9c2dcda7"} Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.136691 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-964564b4f-vkm6n" event={"ID":"0452cb7f-8545-42b9-9456-b29aeaf9c391","Type":"ContainerStarted","Data":"96f7aa2fa61363d0ceebdec6555314b145d68b5e2461518e473f76d3f97ba7fb"} Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.137899 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.142350 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5d114929-218b-465a-9116-ed8144ed0636","Type":"ContainerStarted","Data":"7a08f48c3a2e5031799c652836a2e4db6dbf7eb4a4f1b49abcb9e0575c771574"} Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.186955 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" event={"ID":"c52e4c01-ae05-4a0b-b245-04e570525017","Type":"ContainerStarted","Data":"02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905"} Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.187618 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.242103 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-964564b4f-vkm6n" podStartSLOduration=4.242082388 podStartE2EDuration="4.242082388s" podCreationTimestamp="2025-12-04 06:30:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:18.215453419 +0000 UTC m=+1155.403684194" watchObservedRunningTime="2025-12-04 06:30:18.242082388 +0000 UTC m=+1155.430313163" Dec 04 06:30:18 crc kubenswrapper[4685]: I1204 06:30:18.247186 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" podStartSLOduration=7.247169687 podStartE2EDuration="7.247169687s" podCreationTimestamp="2025-12-04 06:30:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:18.244159513 +0000 UTC m=+1155.432390288" watchObservedRunningTime="2025-12-04 06:30:18.247169687 +0000 UTC m=+1155.435400462" Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.195790 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6985cd56b8-9gnlv" event={"ID":"c657cb75-b17e-41fb-8192-2a6cf3ccbe85","Type":"ContainerStarted","Data":"e2d727cf7329731ea689cc3001d03f5971833f132743fc5beb636e3aca5e23aa"} Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.873178 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.968116 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8daa431f-999a-4213-a745-9d8b03aa028e-config-volume\") pod \"8daa431f-999a-4213-a745-9d8b03aa028e\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.968669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8stp\" (UniqueName: \"kubernetes.io/projected/8daa431f-999a-4213-a745-9d8b03aa028e-kube-api-access-r8stp\") pod \"8daa431f-999a-4213-a745-9d8b03aa028e\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.969421 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8daa431f-999a-4213-a745-9d8b03aa028e-config-volume" (OuterVolumeSpecName: "config-volume") pod "8daa431f-999a-4213-a745-9d8b03aa028e" (UID: "8daa431f-999a-4213-a745-9d8b03aa028e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.978425 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8daa431f-999a-4213-a745-9d8b03aa028e-secret-volume\") pod \"8daa431f-999a-4213-a745-9d8b03aa028e\" (UID: \"8daa431f-999a-4213-a745-9d8b03aa028e\") " Dec 04 06:30:19 crc kubenswrapper[4685]: I1204 06:30:19.986852 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8daa431f-999a-4213-a745-9d8b03aa028e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.093791 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8daa431f-999a-4213-a745-9d8b03aa028e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8daa431f-999a-4213-a745-9d8b03aa028e" (UID: "8daa431f-999a-4213-a745-9d8b03aa028e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.093939 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8daa431f-999a-4213-a745-9d8b03aa028e-kube-api-access-r8stp" (OuterVolumeSpecName: "kube-api-access-r8stp") pod "8daa431f-999a-4213-a745-9d8b03aa028e" (UID: "8daa431f-999a-4213-a745-9d8b03aa028e"). InnerVolumeSpecName "kube-api-access-r8stp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.191716 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8stp\" (UniqueName: \"kubernetes.io/projected/8daa431f-999a-4213-a745-9d8b03aa028e-kube-api-access-r8stp\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.191991 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8daa431f-999a-4213-a745-9d8b03aa028e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.215317 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5d114929-218b-465a-9116-ed8144ed0636","Type":"ContainerStarted","Data":"8a776e43c26649414b0dc2963a340b49b052b633b18c81ea218571f27a43597a"} Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.223774 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.223818 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4" event={"ID":"8daa431f-999a-4213-a745-9d8b03aa028e","Type":"ContainerDied","Data":"38a4c2a6d92a441725360a97988e3ace57b332715413faa7ce883d0c1336a522"} Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.224006 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a4c2a6d92a441725360a97988e3ace57b332715413faa7ce883d0c1336a522" Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.226607 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34","Type":"ContainerStarted","Data":"194e09b1df424174a9f3f618498a0018145f01d76d1843da7d57f62bc9c0f61d"} Dec 04 06:30:20 crc kubenswrapper[4685]: I1204 06:30:20.244626 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.244598944 podStartE2EDuration="10.244598944s" podCreationTimestamp="2025-12-04 06:30:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:20.243925053 +0000 UTC m=+1157.432155828" watchObservedRunningTime="2025-12-04 06:30:20.244598944 +0000 UTC m=+1157.432829719" Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.244674 4685 generic.go:334] "Generic (PLEG): container finished" podID="fd072a3b-1456-43ad-97a7-7326a66a6f9d" containerID="eddd8dea3e1ba7f7c2f7b5188170acda9d987dc8ab31a43c1061e616a86e8b30" exitCode=0 Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.244771 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bljbv" event={"ID":"fd072a3b-1456-43ad-97a7-7326a66a6f9d","Type":"ContainerDied","Data":"eddd8dea3e1ba7f7c2f7b5188170acda9d987dc8ab31a43c1061e616a86e8b30"} Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.258306 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" containerID="dfe3ec87dcc103e750d55dae54a2e0d01f746764027360ddef79b6fa37fbfa75" exitCode=0 Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.258400 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cpkd7" event={"ID":"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3","Type":"ContainerDied","Data":"dfe3ec87dcc103e750d55dae54a2e0d01f746764027360ddef79b6fa37fbfa75"} Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.274171 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6985cd56b8-9gnlv" event={"ID":"c657cb75-b17e-41fb-8192-2a6cf3ccbe85","Type":"ContainerStarted","Data":"bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770"} Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.274227 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6985cd56b8-9gnlv" event={"ID":"c657cb75-b17e-41fb-8192-2a6cf3ccbe85","Type":"ContainerStarted","Data":"7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9"} Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.275069 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.282675 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerStarted","Data":"88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a"} Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.321043 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6985cd56b8-9gnlv" podStartSLOduration=9.321017062 podStartE2EDuration="9.321017062s" podCreationTimestamp="2025-12-04 06:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:21.310871953 +0000 UTC m=+1158.499102728" watchObservedRunningTime="2025-12-04 06:30:21.321017062 +0000 UTC m=+1158.509247837" Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.343808 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.343876 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.386898 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 06:30:21 crc kubenswrapper[4685]: I1204 06:30:21.401797 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.315545 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34","Type":"ContainerStarted","Data":"af51672273a18d10e56be9de171cfdd540c1fb4d658fcd9f511decfd0008f671"} Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.315836 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.315862 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.317514 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.349029 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.349010076999999 podStartE2EDuration="12.349010077s" podCreationTimestamp="2025-12-04 06:30:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:22.332990323 +0000 UTC m=+1159.521221098" watchObservedRunningTime="2025-12-04 06:30:22.349010077 +0000 UTC m=+1159.537240852" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.415117 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-xd9fs"] Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.415397 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerName="dnsmasq-dns" containerID="cri-o://08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b" gracePeriod=10 Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.902535 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.903991 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.904222 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:22 crc kubenswrapper[4685]: I1204 06:30:22.919442 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cpkd7" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.063765 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzvc\" (UniqueName: \"kubernetes.io/projected/fd072a3b-1456-43ad-97a7-7326a66a6f9d-kube-api-access-zmzvc\") pod \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.063873 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-scripts\") pod \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.063915 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-combined-ca-bundle\") pod \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.063999 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-combined-ca-bundle\") pod \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064035 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-credential-keys\") pod \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064118 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-logs\") pod \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064190 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-config-data\") pod \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064219 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-fernet-keys\") pod \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064262 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-config-data\") pod \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064285 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf9kv\" (UniqueName: \"kubernetes.io/projected/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-kube-api-access-rf9kv\") pod \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\" (UID: \"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.064317 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-scripts\") pod \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\" (UID: \"fd072a3b-1456-43ad-97a7-7326a66a6f9d\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.071592 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-logs" (OuterVolumeSpecName: "logs") pod "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" (UID: "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.079182 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.082822 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.086157 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.110827 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-kube-api-access-rf9kv" (OuterVolumeSpecName: "kube-api-access-rf9kv") pod "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" (UID: "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3"). InnerVolumeSpecName "kube-api-access-rf9kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.113326 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-scripts" (OuterVolumeSpecName: "scripts") pod "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" (UID: "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.115597 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fd072a3b-1456-43ad-97a7-7326a66a6f9d" (UID: "fd072a3b-1456-43ad-97a7-7326a66a6f9d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.116711 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-scripts" (OuterVolumeSpecName: "scripts") pod "fd072a3b-1456-43ad-97a7-7326a66a6f9d" (UID: "fd072a3b-1456-43ad-97a7-7326a66a6f9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.119121 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fd072a3b-1456-43ad-97a7-7326a66a6f9d" (UID: "fd072a3b-1456-43ad-97a7-7326a66a6f9d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.126770 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.132062 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd072a3b-1456-43ad-97a7-7326a66a6f9d-kube-api-access-zmzvc" (OuterVolumeSpecName: "kube-api-access-zmzvc") pod "fd072a3b-1456-43ad-97a7-7326a66a6f9d" (UID: "fd072a3b-1456-43ad-97a7-7326a66a6f9d"). InnerVolumeSpecName "kube-api-access-zmzvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.149502 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-config-data" (OuterVolumeSpecName: "config-data") pod "fd072a3b-1456-43ad-97a7-7326a66a6f9d" (UID: "fd072a3b-1456-43ad-97a7-7326a66a6f9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.150116 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" (UID: "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193052 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-nb\") pod \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193104 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-config\") pod \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193165 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-sb\") pod \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193235 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx8tj\" (UniqueName: \"kubernetes.io/projected/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-kube-api-access-gx8tj\") pod \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193357 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-swift-storage-0\") pod \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193461 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-svc\") pod \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\" (UID: \"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7\") " Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193903 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193915 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzvc\" (UniqueName: \"kubernetes.io/projected/fd072a3b-1456-43ad-97a7-7326a66a6f9d-kube-api-access-zmzvc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193927 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193935 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193944 4685 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193952 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193960 4685 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.193968 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf9kv\" (UniqueName: \"kubernetes.io/projected/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-kube-api-access-rf9kv\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.235596 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd072a3b-1456-43ad-97a7-7326a66a6f9d" (UID: "fd072a3b-1456-43ad-97a7-7326a66a6f9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.260586 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-config-data" (OuterVolumeSpecName: "config-data") pod "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" (UID: "6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.271617 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-kube-api-access-gx8tj" (OuterVolumeSpecName: "kube-api-access-gx8tj") pod "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" (UID: "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7"). InnerVolumeSpecName "kube-api-access-gx8tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.299667 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.299702 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd072a3b-1456-43ad-97a7-7326a66a6f9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.299713 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx8tj\" (UniqueName: \"kubernetes.io/projected/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-kube-api-access-gx8tj\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.399163 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bljbv" event={"ID":"fd072a3b-1456-43ad-97a7-7326a66a6f9d","Type":"ContainerDied","Data":"bada0e63685a17346e53af214b5c758197e63770fd8d84dc7c499c0e78bf4d10"} Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.399217 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bada0e63685a17346e53af214b5c758197e63770fd8d84dc7c499c0e78bf4d10" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.399327 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bljbv" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.401149 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-config" (OuterVolumeSpecName: "config") pod "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" (UID: "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.401904 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" (UID: "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.426993 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" (UID: "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.446532 4685 generic.go:334] "Generic (PLEG): container finished" podID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerID="08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b" exitCode=0 Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.446636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" event={"ID":"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7","Type":"ContainerDied","Data":"08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b"} Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.446679 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" event={"ID":"26ff3a03-81ee-4d98-a6bc-829bea2ed9f7","Type":"ContainerDied","Data":"355451359a7464821e403bca6e4cb1703fdc5401d7563269c57f16efce0e4dda"} Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.446697 4685 scope.go:117] "RemoveContainer" containerID="08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.446850 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-xd9fs" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.469377 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-679cdf976b-2nlz9"] Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478027 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="init" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478055 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="init" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478075 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerName="init" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478081 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerName="init" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478089 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerName="dnsmasq-dns" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478095 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerName="dnsmasq-dns" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478112 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd072a3b-1456-43ad-97a7-7326a66a6f9d" containerName="keystone-bootstrap" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478120 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd072a3b-1456-43ad-97a7-7326a66a6f9d" containerName="keystone-bootstrap" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478130 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478136 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478153 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" containerName="placement-db-sync" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478159 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" containerName="placement-db-sync" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.478171 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8daa431f-999a-4213-a745-9d8b03aa028e" containerName="collect-profiles" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478177 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8daa431f-999a-4213-a745-9d8b03aa028e" containerName="collect-profiles" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478338 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" containerName="placement-db-sync" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478358 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="991d5443-6323-48ba-9a55-832b4d80820b" containerName="dnsmasq-dns" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478368 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" containerName="dnsmasq-dns" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478375 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8daa431f-999a-4213-a745-9d8b03aa028e" containerName="collect-profiles" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478386 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd072a3b-1456-43ad-97a7-7326a66a6f9d" containerName="keystone-bootstrap" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478955 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cpkd7" event={"ID":"6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3","Type":"ContainerDied","Data":"5bd8e4476691b03027da40d0743abc8e8b9f857c3f6626d33332a7e84189d72a"} Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.478995 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bd8e4476691b03027da40d0743abc8e8b9f857c3f6626d33332a7e84189d72a" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.479085 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.471567 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cpkd7" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.483023 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" (UID: "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.490831 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.491193 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.491330 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.491465 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c6l2p" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.491462 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.491999 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.492962 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" (UID: "26ff3a03-81ee-4d98-a6bc-829bea2ed9f7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.502691 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.502719 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.502728 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.502736 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.502746 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.510144 4685 scope.go:117] "RemoveContainer" containerID="dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.528389 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-679cdf976b-2nlz9"] Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.563900 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6bc7689ffb-9thrj"] Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.565695 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.567381 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6bc7689ffb-9thrj"] Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.576003 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.576209 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.576326 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ccdtn" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.576554 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.576671 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708543 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-combined-ca-bundle\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708585 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-fernet-keys\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708617 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2tzw\" (UniqueName: \"kubernetes.io/projected/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-kube-api-access-j2tzw\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708650 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-scripts\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708678 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-credential-keys\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708696 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-public-tls-certs\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708713 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-config-data\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708730 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-logs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708746 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-internal-tls-certs\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708777 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-scripts\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708807 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-public-tls-certs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708835 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-combined-ca-bundle\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708874 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfdw8\" (UniqueName: \"kubernetes.io/projected/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-kube-api-access-vfdw8\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708891 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-internal-tls-certs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.708904 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-config-data\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.810989 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-combined-ca-bundle\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811267 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-fernet-keys\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811363 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2tzw\" (UniqueName: \"kubernetes.io/projected/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-kube-api-access-j2tzw\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811457 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-scripts\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811537 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-credential-keys\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811611 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-public-tls-certs\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811686 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-config-data\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.811896 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-logs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.812362 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-internal-tls-certs\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.812540 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-scripts\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.813111 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-public-tls-certs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.813227 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-combined-ca-bundle\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.814151 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfdw8\" (UniqueName: \"kubernetes.io/projected/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-kube-api-access-vfdw8\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.814302 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-internal-tls-certs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.814397 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-config-data\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.816921 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-logs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.822428 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-fernet-keys\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.825438 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-credential-keys\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.825835 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-internal-tls-certs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.831768 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-combined-ca-bundle\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.834336 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-scripts\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.834528 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-combined-ca-bundle\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.834977 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-public-tls-certs\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.835666 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-config-data\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.835682 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-config-data\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.835954 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-scripts\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.839911 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-public-tls-certs\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.840122 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfdw8\" (UniqueName: \"kubernetes.io/projected/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-kube-api-access-vfdw8\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.840271 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2tzw\" (UniqueName: \"kubernetes.io/projected/4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55-kube-api-access-j2tzw\") pod \"placement-6bc7689ffb-9thrj\" (UID: \"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55\") " pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.843886 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd-internal-tls-certs\") pod \"keystone-679cdf976b-2nlz9\" (UID: \"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd\") " pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.906590 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.962553 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-xd9fs"] Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.962561 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.965622 4685 scope.go:117] "RemoveContainer" containerID="08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.971684 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b\": container with ID starting with 08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b not found: ID does not exist" containerID="08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.971726 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b"} err="failed to get container status \"08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b\": rpc error: code = NotFound desc = could not find container \"08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b\": container with ID starting with 08e532597cf7405249f39c6c332ce3cd06d16f1db8527c9f07822c103725c09b not found: ID does not exist" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.971761 4685 scope.go:117] "RemoveContainer" containerID="dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486" Dec 04 06:30:23 crc kubenswrapper[4685]: E1204 06:30:23.975572 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486\": container with ID starting with dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486 not found: ID does not exist" containerID="dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.975603 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486"} err="failed to get container status \"dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486\": rpc error: code = NotFound desc = could not find container \"dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486\": container with ID starting with dda48952615ae91f8e28a4e11447352a09dc3e0d48a79c332ce6018bbac2b486 not found: ID does not exist" Dec 04 06:30:23 crc kubenswrapper[4685]: I1204 06:30:23.982254 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-xd9fs"] Dec 04 06:30:24 crc kubenswrapper[4685]: I1204 06:30:24.511190 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5498d" event={"ID":"6f6bce91-0f30-4543-ac33-b34877206b05","Type":"ContainerStarted","Data":"17385e95af7a11a08f64971163a7c1d699257f4e8d61af5368dbd211eb6c9070"} Dec 04 06:30:24 crc kubenswrapper[4685]: I1204 06:30:24.520487 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:30:24 crc kubenswrapper[4685]: I1204 06:30:24.536736 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5498d" podStartSLOduration=2.884098805 podStartE2EDuration="41.536709049s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="2025-12-04 06:29:45.199933529 +0000 UTC m=+1122.388164314" lastFinishedPulling="2025-12-04 06:30:23.852543783 +0000 UTC m=+1161.040774558" observedRunningTime="2025-12-04 06:30:24.528886463 +0000 UTC m=+1161.717117238" watchObservedRunningTime="2025-12-04 06:30:24.536709049 +0000 UTC m=+1161.724939824" Dec 04 06:30:24 crc kubenswrapper[4685]: W1204 06:30:24.545061 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3a0ad7a_fa07_41f5_8fa7_d474d8b052fd.slice/crio-ed36803753288edcf97afd542ac558d28cbf13663cf2f13e92df0d79bf2711ee WatchSource:0}: Error finding container ed36803753288edcf97afd542ac558d28cbf13663cf2f13e92df0d79bf2711ee: Status 404 returned error can't find the container with id ed36803753288edcf97afd542ac558d28cbf13663cf2f13e92df0d79bf2711ee Dec 04 06:30:24 crc kubenswrapper[4685]: I1204 06:30:24.564208 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-679cdf976b-2nlz9"] Dec 04 06:30:24 crc kubenswrapper[4685]: I1204 06:30:24.677510 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6bc7689ffb-9thrj"] Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.197247 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ff3a03-81ee-4d98-a6bc-829bea2ed9f7" path="/var/lib/kubelet/pods/26ff3a03-81ee-4d98-a6bc-829bea2ed9f7/volumes" Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.535271 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6bc7689ffb-9thrj" event={"ID":"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55","Type":"ContainerStarted","Data":"7264ce9199eee2ec6a2ce7b7c48074fda341766adeb4df41f177b043e6183242"} Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.535753 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.535787 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6bc7689ffb-9thrj" event={"ID":"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55","Type":"ContainerStarted","Data":"3b9ed83c6909ce99f47ef90ee847df1e010861684089a07633441682ebecccac"} Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.535802 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6bc7689ffb-9thrj" event={"ID":"4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55","Type":"ContainerStarted","Data":"b0478a308cb24737ad43c9712756294a5a2267e40f73380a5b9e4943d5980a09"} Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.546388 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-679cdf976b-2nlz9" event={"ID":"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd","Type":"ContainerStarted","Data":"5a96ab2a7e8a7a0cb12d6af0b651720aab320244e2973abf240334d7a2d965b0"} Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.546478 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-679cdf976b-2nlz9" event={"ID":"c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd","Type":"ContainerStarted","Data":"ed36803753288edcf97afd542ac558d28cbf13663cf2f13e92df0d79bf2711ee"} Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.546640 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.590385 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6bc7689ffb-9thrj" podStartSLOduration=2.590362942 podStartE2EDuration="2.590362942s" podCreationTimestamp="2025-12-04 06:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:25.567589825 +0000 UTC m=+1162.755820610" watchObservedRunningTime="2025-12-04 06:30:25.590362942 +0000 UTC m=+1162.778593717" Dec 04 06:30:25 crc kubenswrapper[4685]: I1204 06:30:25.597239 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-679cdf976b-2nlz9" podStartSLOduration=2.597219598 podStartE2EDuration="2.597219598s" podCreationTimestamp="2025-12-04 06:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:25.584069573 +0000 UTC m=+1162.772300348" watchObservedRunningTime="2025-12-04 06:30:25.597219598 +0000 UTC m=+1162.785450393" Dec 04 06:30:26 crc kubenswrapper[4685]: I1204 06:30:26.255601 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 06:30:26 crc kubenswrapper[4685]: I1204 06:30:26.557152 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:27 crc kubenswrapper[4685]: I1204 06:30:27.148607 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 06:30:27 crc kubenswrapper[4685]: I1204 06:30:27.570326 4685 generic.go:334] "Generic (PLEG): container finished" podID="6f6bce91-0f30-4543-ac33-b34877206b05" containerID="17385e95af7a11a08f64971163a7c1d699257f4e8d61af5368dbd211eb6c9070" exitCode=0 Dec 04 06:30:27 crc kubenswrapper[4685]: I1204 06:30:27.570658 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5498d" event={"ID":"6f6bce91-0f30-4543-ac33-b34877206b05","Type":"ContainerDied","Data":"17385e95af7a11a08f64971163a7c1d699257f4e8d61af5368dbd211eb6c9070"} Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.542345 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5498d" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.617300 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5498d" event={"ID":"6f6bce91-0f30-4543-ac33-b34877206b05","Type":"ContainerDied","Data":"6611159eddc7c9c87ead02365070f6389b0e9507202bb41edb1ff7681210b158"} Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.617352 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6611159eddc7c9c87ead02365070f6389b0e9507202bb41edb1ff7681210b158" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.617432 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5498d" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.701594 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-db-sync-config-data\") pod \"6f6bce91-0f30-4543-ac33-b34877206b05\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.701677 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-combined-ca-bundle\") pod \"6f6bce91-0f30-4543-ac33-b34877206b05\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.701725 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48pl6\" (UniqueName: \"kubernetes.io/projected/6f6bce91-0f30-4543-ac33-b34877206b05-kube-api-access-48pl6\") pod \"6f6bce91-0f30-4543-ac33-b34877206b05\" (UID: \"6f6bce91-0f30-4543-ac33-b34877206b05\") " Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.705736 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6f6bce91-0f30-4543-ac33-b34877206b05" (UID: "6f6bce91-0f30-4543-ac33-b34877206b05"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.706504 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f6bce91-0f30-4543-ac33-b34877206b05-kube-api-access-48pl6" (OuterVolumeSpecName: "kube-api-access-48pl6") pod "6f6bce91-0f30-4543-ac33-b34877206b05" (UID: "6f6bce91-0f30-4543-ac33-b34877206b05"). InnerVolumeSpecName "kube-api-access-48pl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.730835 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f6bce91-0f30-4543-ac33-b34877206b05" (UID: "6f6bce91-0f30-4543-ac33-b34877206b05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.803773 4685 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.804061 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6bce91-0f30-4543-ac33-b34877206b05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:30 crc kubenswrapper[4685]: I1204 06:30:30.804073 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48pl6\" (UniqueName: \"kubernetes.io/projected/6f6bce91-0f30-4543-ac33-b34877206b05-kube-api-access-48pl6\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:31 crc kubenswrapper[4685]: E1204 06:30:31.071024 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f6bce91_0f30_4543_ac33_b34877206b05.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f6bce91_0f30_4543_ac33_b34877206b05.slice/crio-6611159eddc7c9c87ead02365070f6389b0e9507202bb41edb1ff7681210b158\": RecentStats: unable to find data in memory cache]" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.353189 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.353262 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.389688 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.398611 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.629239 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerStarted","Data":"99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a"} Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.635129 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-59cdh" event={"ID":"964316c7-6da5-4725-8140-6d249ee01a68","Type":"ContainerStarted","Data":"b8e09f16dce9dcdec259fcc9b33007c3a656dcfd4a71640869312ea4072637dc"} Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.635265 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.635290 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.659313 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-59cdh" podStartSLOduration=3.570774931 podStartE2EDuration="48.659290982s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="2025-12-04 06:29:45.30769631 +0000 UTC m=+1122.495927085" lastFinishedPulling="2025-12-04 06:30:30.396212341 +0000 UTC m=+1167.584443136" observedRunningTime="2025-12-04 06:30:31.656203015 +0000 UTC m=+1168.844433800" watchObservedRunningTime="2025-12-04 06:30:31.659290982 +0000 UTC m=+1168.847521777" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.877489 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-84f9d55fbc-qkcbh"] Dec 04 06:30:31 crc kubenswrapper[4685]: E1204 06:30:31.877943 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f6bce91-0f30-4543-ac33-b34877206b05" containerName="barbican-db-sync" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.877959 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f6bce91-0f30-4543-ac33-b34877206b05" containerName="barbican-db-sync" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.878119 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f6bce91-0f30-4543-ac33-b34877206b05" containerName="barbican-db-sync" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.879118 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.890468 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.890822 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-p7d29" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.890995 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.893616 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84f9d55fbc-qkcbh"] Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.954828 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-64bd497db6-z9fft"] Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.958501 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.960199 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64bd497db6-z9fft"] Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.969856 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.974836 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-ssdgf"] Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.996226 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-ssdgf"] Dec 04 06:30:31 crc kubenswrapper[4685]: I1204 06:30:31.996282 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.030712 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-logs\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.030831 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-combined-ca-bundle\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.030871 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-config-data-custom\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.030945 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-config-data\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.030975 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-config-data\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.031004 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gksj\" (UniqueName: \"kubernetes.io/projected/a1d37923-dce0-47f0-990f-92efc1754323-kube-api-access-8gksj\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.031040 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-config-data-custom\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.031055 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1d37923-dce0-47f0-990f-92efc1754323-logs\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.031078 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmf4x\" (UniqueName: \"kubernetes.io/projected/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-kube-api-access-nmf4x\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.031107 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-combined-ca-bundle\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.132893 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-combined-ca-bundle\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133162 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133214 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-logs\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133274 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-combined-ca-bundle\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133303 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-config-data-custom\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133325 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133359 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133399 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-config-data\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133467 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-config-data\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133483 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-config\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133505 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gksj\" (UniqueName: \"kubernetes.io/projected/a1d37923-dce0-47f0-990f-92efc1754323-kube-api-access-8gksj\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133522 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrhgm\" (UniqueName: \"kubernetes.io/projected/17281897-9381-4ab1-a79f-9cd39a8b5508-kube-api-access-vrhgm\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133547 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133569 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-config-data-custom\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133585 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1d37923-dce0-47f0-990f-92efc1754323-logs\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.133610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmf4x\" (UniqueName: \"kubernetes.io/projected/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-kube-api-access-nmf4x\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.145114 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-combined-ca-bundle\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.145457 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c7d4bcdc6-psn87"] Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.146947 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.148115 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-logs\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.148555 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-config-data\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.149744 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.151940 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-config-data\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.153766 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1d37923-dce0-47f0-990f-92efc1754323-logs\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.156334 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1d37923-dce0-47f0-990f-92efc1754323-config-data-custom\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.158691 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-config-data-custom\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.167981 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-combined-ca-bundle\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.168507 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmf4x\" (UniqueName: \"kubernetes.io/projected/1d7a178f-f4ba-4a18-ab97-f2d256e1edc1-kube-api-access-nmf4x\") pod \"barbican-keystone-listener-64bd497db6-z9fft\" (UID: \"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1\") " pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.173370 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7d4bcdc6-psn87"] Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.179644 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gksj\" (UniqueName: \"kubernetes.io/projected/a1d37923-dce0-47f0-990f-92efc1754323-kube-api-access-8gksj\") pod \"barbican-worker-84f9d55fbc-qkcbh\" (UID: \"a1d37923-dce0-47f0-990f-92efc1754323\") " pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.215836 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84f9d55fbc-qkcbh" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.239901 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.239958 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-combined-ca-bundle\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240084 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data-custom\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240160 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240217 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240252 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-config\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240289 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrhgm\" (UniqueName: \"kubernetes.io/projected/17281897-9381-4ab1-a79f-9cd39a8b5508-kube-api-access-vrhgm\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240317 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tbz2\" (UniqueName: \"kubernetes.io/projected/9d68936d-d2fb-489f-91e3-de59fce181a2-kube-api-access-8tbz2\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240338 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240368 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d68936d-d2fb-489f-91e3-de59fce181a2-logs\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.240392 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.241460 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.244147 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.246654 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.252844 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-config\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.263026 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.275529 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrhgm\" (UniqueName: \"kubernetes.io/projected/17281897-9381-4ab1-a79f-9cd39a8b5508-kube-api-access-vrhgm\") pod \"dnsmasq-dns-848cf88cfc-ssdgf\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.323981 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.367034 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.369131 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data-custom\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.369266 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tbz2\" (UniqueName: \"kubernetes.io/projected/9d68936d-d2fb-489f-91e3-de59fce181a2-kube-api-access-8tbz2\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.369309 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d68936d-d2fb-489f-91e3-de59fce181a2-logs\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.369343 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.369374 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-combined-ca-bundle\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.369876 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d68936d-d2fb-489f-91e3-de59fce181a2-logs\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.375086 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-combined-ca-bundle\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.377501 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.385789 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data-custom\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.385851 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tbz2\" (UniqueName: \"kubernetes.io/projected/9d68936d-d2fb-489f-91e3-de59fce181a2-kube-api-access-8tbz2\") pod \"barbican-api-6c7d4bcdc6-psn87\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.519868 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.852620 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84f9d55fbc-qkcbh"] Dec 04 06:30:32 crc kubenswrapper[4685]: I1204 06:30:32.909117 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bdfd7d6f6-8slln" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.081836 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c9548d9dd-h5vqf" podUID="0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.185392 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64bd497db6-z9fft"] Dec 04 06:30:33 crc kubenswrapper[4685]: W1204 06:30:33.196700 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d7a178f_f4ba_4a18_ab97_f2d256e1edc1.slice/crio-2d660f85f48deee749be7e9b89cc0b5cfd8eb1fb0864ceb8ba157533cd2f9327 WatchSource:0}: Error finding container 2d660f85f48deee749be7e9b89cc0b5cfd8eb1fb0864ceb8ba157533cd2f9327: Status 404 returned error can't find the container with id 2d660f85f48deee749be7e9b89cc0b5cfd8eb1fb0864ceb8ba157533cd2f9327 Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.269105 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-ssdgf"] Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.355216 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7d4bcdc6-psn87"] Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.706432 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84f9d55fbc-qkcbh" event={"ID":"a1d37923-dce0-47f0-990f-92efc1754323","Type":"ContainerStarted","Data":"31a2378842d35a76b063718439228db8c2608ebf082f7f905ac2b7731d92cedc"} Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.716198 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" event={"ID":"17281897-9381-4ab1-a79f-9cd39a8b5508","Type":"ContainerStarted","Data":"51580cfdd26d0ca0fa7e8e5dc00d9c650bbc187e1febbba176a008b448385186"} Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.725155 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7d4bcdc6-psn87" event={"ID":"9d68936d-d2fb-489f-91e3-de59fce181a2","Type":"ContainerStarted","Data":"79e29e3e7a5584aca2d3e0f8c3e1c7b191194abe8456e1722680833d9ff2fb65"} Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.736290 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.736315 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:30:33 crc kubenswrapper[4685]: I1204 06:30:33.736919 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" event={"ID":"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1","Type":"ContainerStarted","Data":"2d660f85f48deee749be7e9b89cc0b5cfd8eb1fb0864ceb8ba157533cd2f9327"} Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.762512 4685 generic.go:334] "Generic (PLEG): container finished" podID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerID="296cb7cbca21efad93ddd8a273033cd984e544b2321dadb75a686d630659a213" exitCode=0 Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.763045 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" event={"ID":"17281897-9381-4ab1-a79f-9cd39a8b5508","Type":"ContainerDied","Data":"296cb7cbca21efad93ddd8a273033cd984e544b2321dadb75a686d630659a213"} Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.769460 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7d4bcdc6-psn87" event={"ID":"9d68936d-d2fb-489f-91e3-de59fce181a2","Type":"ContainerStarted","Data":"e14dd9c4bcc207cf304c068d5f2b25174cc34ed3a0111a84d8dd8a22ecac6799"} Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.769516 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7d4bcdc6-psn87" event={"ID":"9d68936d-d2fb-489f-91e3-de59fce181a2","Type":"ContainerStarted","Data":"e5811deb4eb8030ae26f6904e5f133b7306db0b535f20eec60092c49123ba493"} Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.770536 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.770575 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.816223 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c7d4bcdc6-psn87" podStartSLOduration=2.816200011 podStartE2EDuration="2.816200011s" podCreationTimestamp="2025-12-04 06:30:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:34.813053812 +0000 UTC m=+1172.001284587" watchObservedRunningTime="2025-12-04 06:30:34.816200011 +0000 UTC m=+1172.004430786" Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.973154 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-b6b84b88b-bjcmm"] Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.974782 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.987609 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.988147 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 04 06:30:34 crc kubenswrapper[4685]: I1204 06:30:34.997435 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b6b84b88b-bjcmm"] Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-config-data-custom\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042159 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-internal-tls-certs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042188 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-public-tls-certs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042206 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-combined-ca-bundle\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042236 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66becbc7-a80f-45d5-a711-f0c50c304034-logs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042269 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h9fd\" (UniqueName: \"kubernetes.io/projected/66becbc7-a80f-45d5-a711-f0c50c304034-kube-api-access-6h9fd\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.042438 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-config-data\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.145878 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66becbc7-a80f-45d5-a711-f0c50c304034-logs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.145951 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h9fd\" (UniqueName: \"kubernetes.io/projected/66becbc7-a80f-45d5-a711-f0c50c304034-kube-api-access-6h9fd\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.146045 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-config-data\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.146127 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-config-data-custom\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.146160 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-internal-tls-certs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.146189 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-public-tls-certs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.146217 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-combined-ca-bundle\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.146343 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66becbc7-a80f-45d5-a711-f0c50c304034-logs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.157021 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-combined-ca-bundle\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.157855 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-public-tls-certs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.159988 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-internal-tls-certs\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.160649 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-config-data-custom\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.162373 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66becbc7-a80f-45d5-a711-f0c50c304034-config-data\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.182887 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h9fd\" (UniqueName: \"kubernetes.io/projected/66becbc7-a80f-45d5-a711-f0c50c304034-kube-api-access-6h9fd\") pod \"barbican-api-b6b84b88b-bjcmm\" (UID: \"66becbc7-a80f-45d5-a711-f0c50c304034\") " pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.298822 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.388028 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.388179 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:30:35 crc kubenswrapper[4685]: I1204 06:30:35.393932 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 06:30:36 crc kubenswrapper[4685]: I1204 06:30:36.731029 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b6b84b88b-bjcmm"] Dec 04 06:30:36 crc kubenswrapper[4685]: W1204 06:30:36.783229 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66becbc7_a80f_45d5_a711_f0c50c304034.slice/crio-3c28eca7c8ff128915a7cb0dfab7024610f60369a32f7188a74c74705c5d73bd WatchSource:0}: Error finding container 3c28eca7c8ff128915a7cb0dfab7024610f60369a32f7188a74c74705c5d73bd: Status 404 returned error can't find the container with id 3c28eca7c8ff128915a7cb0dfab7024610f60369a32f7188a74c74705c5d73bd Dec 04 06:30:36 crc kubenswrapper[4685]: I1204 06:30:36.798641 4685 generic.go:334] "Generic (PLEG): container finished" podID="964316c7-6da5-4725-8140-6d249ee01a68" containerID="b8e09f16dce9dcdec259fcc9b33007c3a656dcfd4a71640869312ea4072637dc" exitCode=0 Dec 04 06:30:36 crc kubenswrapper[4685]: I1204 06:30:36.798719 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-59cdh" event={"ID":"964316c7-6da5-4725-8140-6d249ee01a68","Type":"ContainerDied","Data":"b8e09f16dce9dcdec259fcc9b33007c3a656dcfd4a71640869312ea4072637dc"} Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.828955 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84f9d55fbc-qkcbh" event={"ID":"a1d37923-dce0-47f0-990f-92efc1754323","Type":"ContainerStarted","Data":"515e1c26d30a10329f9d6e754bef970bdaf1a0579c155c25c6604c00e13752ef"} Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.830832 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" event={"ID":"17281897-9381-4ab1-a79f-9cd39a8b5508","Type":"ContainerStarted","Data":"3e3bd003cb39deea54c92c9d70e098dacbe512bcc173de0ea794de690f77dea4"} Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.831507 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.832717 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b6b84b88b-bjcmm" event={"ID":"66becbc7-a80f-45d5-a711-f0c50c304034","Type":"ContainerStarted","Data":"5ade8188a93e2a08c4b2a50ac64d464866248a2df3083c0874cd6f385ecdba48"} Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.832758 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b6b84b88b-bjcmm" event={"ID":"66becbc7-a80f-45d5-a711-f0c50c304034","Type":"ContainerStarted","Data":"3c28eca7c8ff128915a7cb0dfab7024610f60369a32f7188a74c74705c5d73bd"} Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.842909 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" event={"ID":"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1","Type":"ContainerStarted","Data":"989ae327acf8c1abcc71dc58616848a36a3988053d23228ad9b5aecaf70e71d8"} Dec 04 06:30:37 crc kubenswrapper[4685]: I1204 06:30:37.859252 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" podStartSLOduration=6.859224756 podStartE2EDuration="6.859224756s" podCreationTimestamp="2025-12-04 06:30:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:37.857191711 +0000 UTC m=+1175.045422526" watchObservedRunningTime="2025-12-04 06:30:37.859224756 +0000 UTC m=+1175.047455531" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.034270 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-59cdh" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.104522 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-scripts\") pod \"964316c7-6da5-4725-8140-6d249ee01a68\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.104619 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/964316c7-6da5-4725-8140-6d249ee01a68-etc-machine-id\") pod \"964316c7-6da5-4725-8140-6d249ee01a68\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.104650 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-db-sync-config-data\") pod \"964316c7-6da5-4725-8140-6d249ee01a68\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.104828 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psn2s\" (UniqueName: \"kubernetes.io/projected/964316c7-6da5-4725-8140-6d249ee01a68-kube-api-access-psn2s\") pod \"964316c7-6da5-4725-8140-6d249ee01a68\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.104853 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-config-data\") pod \"964316c7-6da5-4725-8140-6d249ee01a68\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.104883 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-combined-ca-bundle\") pod \"964316c7-6da5-4725-8140-6d249ee01a68\" (UID: \"964316c7-6da5-4725-8140-6d249ee01a68\") " Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.105037 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/964316c7-6da5-4725-8140-6d249ee01a68-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "964316c7-6da5-4725-8140-6d249ee01a68" (UID: "964316c7-6da5-4725-8140-6d249ee01a68"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.105379 4685 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/964316c7-6da5-4725-8140-6d249ee01a68-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.112527 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "964316c7-6da5-4725-8140-6d249ee01a68" (UID: "964316c7-6da5-4725-8140-6d249ee01a68"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.112680 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-scripts" (OuterVolumeSpecName: "scripts") pod "964316c7-6da5-4725-8140-6d249ee01a68" (UID: "964316c7-6da5-4725-8140-6d249ee01a68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.118759 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964316c7-6da5-4725-8140-6d249ee01a68-kube-api-access-psn2s" (OuterVolumeSpecName: "kube-api-access-psn2s") pod "964316c7-6da5-4725-8140-6d249ee01a68" (UID: "964316c7-6da5-4725-8140-6d249ee01a68"). InnerVolumeSpecName "kube-api-access-psn2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.139344 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "964316c7-6da5-4725-8140-6d249ee01a68" (UID: "964316c7-6da5-4725-8140-6d249ee01a68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.170808 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-config-data" (OuterVolumeSpecName: "config-data") pod "964316c7-6da5-4725-8140-6d249ee01a68" (UID: "964316c7-6da5-4725-8140-6d249ee01a68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.207320 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.207355 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.207364 4685 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.207375 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psn2s\" (UniqueName: \"kubernetes.io/projected/964316c7-6da5-4725-8140-6d249ee01a68-kube-api-access-psn2s\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.207386 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964316c7-6da5-4725-8140-6d249ee01a68-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.369855 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.448102 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-tvq7r"] Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.448374 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" containerName="dnsmasq-dns" containerID="cri-o://02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905" gracePeriod=10 Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.477380 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.942726 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-59cdh" event={"ID":"964316c7-6da5-4725-8140-6d249ee01a68","Type":"ContainerDied","Data":"a239456b7df1c8a7bc0119a81e155fe9044fe10c2954e7d31b8149eccb6b244f"} Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.943138 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a239456b7df1c8a7bc0119a81e155fe9044fe10c2954e7d31b8149eccb6b244f" Dec 04 06:30:42 crc kubenswrapper[4685]: I1204 06:30:42.942829 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-59cdh" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.400746 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:43 crc kubenswrapper[4685]: E1204 06:30:43.402888 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964316c7-6da5-4725-8140-6d249ee01a68" containerName="cinder-db-sync" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.402955 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="964316c7-6da5-4725-8140-6d249ee01a68" containerName="cinder-db-sync" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.403613 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="964316c7-6da5-4725-8140-6d249ee01a68" containerName="cinder-db-sync" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.408788 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.415834 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.415930 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4nxjd" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.415848 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.416101 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.475016 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502281 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-2g4gj"] Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502658 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-scripts\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502715 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502818 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502854 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502886 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfksm\" (UniqueName: \"kubernetes.io/projected/b690f179-121b-48b5-87bc-16d8c9dfb89e-kube-api-access-xfksm\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.502997 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b690f179-121b-48b5-87bc-16d8c9dfb89e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.507914 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.531476 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-2g4gj"] Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604395 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604480 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604537 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b690f179-121b-48b5-87bc-16d8c9dfb89e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604578 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k6tz\" (UniqueName: \"kubernetes.io/projected/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-kube-api-access-2k6tz\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604625 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-scripts\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604646 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604711 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-svc\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-config\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604762 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604801 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604826 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.604854 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfksm\" (UniqueName: \"kubernetes.io/projected/b690f179-121b-48b5-87bc-16d8c9dfb89e-kube-api-access-xfksm\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.605374 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b690f179-121b-48b5-87bc-16d8c9dfb89e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.610803 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-scripts\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.614774 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.619133 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.622242 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.622596 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfksm\" (UniqueName: \"kubernetes.io/projected/b690f179-121b-48b5-87bc-16d8c9dfb89e-kube-api-access-xfksm\") pod \"cinder-scheduler-0\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.629202 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.630984 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.649628 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.658519 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.706031 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k6tz\" (UniqueName: \"kubernetes.io/projected/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-kube-api-access-2k6tz\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.706128 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-svc\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.706155 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-config\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.706213 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.706284 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.706308 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.707279 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.708218 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.708587 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-svc\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.708997 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-config\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.709795 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.741902 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k6tz\" (UniqueName: \"kubernetes.io/projected/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-kube-api-access-2k6tz\") pod \"dnsmasq-dns-6578955fd5-2g4gj\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.787887 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.791830 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.809772 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.809945 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data-custom\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.810607 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.810715 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wtdm\" (UniqueName: \"kubernetes.io/projected/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-kube-api-access-9wtdm\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.813371 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.813466 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-logs\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.813531 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-scripts\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.842050 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.914872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-sb\") pod \"c52e4c01-ae05-4a0b-b245-04e570525017\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915047 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-config\") pod \"c52e4c01-ae05-4a0b-b245-04e570525017\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915153 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-nb\") pod \"c52e4c01-ae05-4a0b-b245-04e570525017\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915197 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zz5b\" (UniqueName: \"kubernetes.io/projected/c52e4c01-ae05-4a0b-b245-04e570525017-kube-api-access-2zz5b\") pod \"c52e4c01-ae05-4a0b-b245-04e570525017\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915245 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-svc\") pod \"c52e4c01-ae05-4a0b-b245-04e570525017\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915326 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-swift-storage-0\") pod \"c52e4c01-ae05-4a0b-b245-04e570525017\" (UID: \"c52e4c01-ae05-4a0b-b245-04e570525017\") " Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915553 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wtdm\" (UniqueName: \"kubernetes.io/projected/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-kube-api-access-9wtdm\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915593 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915619 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-logs\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915642 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-scripts\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915753 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915794 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data-custom\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915817 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.915911 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.921932 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-logs\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.953689 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c52e4c01-ae05-4a0b-b245-04e570525017-kube-api-access-2zz5b" (OuterVolumeSpecName: "kube-api-access-2zz5b") pod "c52e4c01-ae05-4a0b-b245-04e570525017" (UID: "c52e4c01-ae05-4a0b-b245-04e570525017"). InnerVolumeSpecName "kube-api-access-2zz5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.962243 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.963374 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.964323 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data-custom\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.970160 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-scripts\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:43 crc kubenswrapper[4685]: I1204 06:30:43.972292 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wtdm\" (UniqueName: \"kubernetes.io/projected/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-kube-api-access-9wtdm\") pod \"cinder-api-0\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " pod="openstack/cinder-api-0" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.010096 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.011039 4685 generic.go:334] "Generic (PLEG): container finished" podID="c52e4c01-ae05-4a0b-b245-04e570525017" containerID="02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905" exitCode=0 Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.011077 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" event={"ID":"c52e4c01-ae05-4a0b-b245-04e570525017","Type":"ContainerDied","Data":"02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905"} Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.011106 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" event={"ID":"c52e4c01-ae05-4a0b-b245-04e570525017","Type":"ContainerDied","Data":"74144e9ce8665bade7060dca11978b27a2bc5c59b9bd259caa8f26ee2f42bd52"} Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.011126 4685 scope.go:117] "RemoveContainer" containerID="02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.011302 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-tvq7r" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.021074 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zz5b\" (UniqueName: \"kubernetes.io/projected/c52e4c01-ae05-4a0b-b245-04e570525017-kube-api-access-2zz5b\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.125291 4685 scope.go:117] "RemoveContainer" containerID="6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.212807 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c52e4c01-ae05-4a0b-b245-04e570525017" (UID: "c52e4c01-ae05-4a0b-b245-04e570525017"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.219219 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c52e4c01-ae05-4a0b-b245-04e570525017" (UID: "c52e4c01-ae05-4a0b-b245-04e570525017"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.227656 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.227682 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.242628 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c52e4c01-ae05-4a0b-b245-04e570525017" (UID: "c52e4c01-ae05-4a0b-b245-04e570525017"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.254657 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-config" (OuterVolumeSpecName: "config") pod "c52e4c01-ae05-4a0b-b245-04e570525017" (UID: "c52e4c01-ae05-4a0b-b245-04e570525017"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.278998 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c52e4c01-ae05-4a0b-b245-04e570525017" (UID: "c52e4c01-ae05-4a0b-b245-04e570525017"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.329683 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.329721 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.329733 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c52e4c01-ae05-4a0b-b245-04e570525017-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.367393 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.393544 4685 scope.go:117] "RemoveContainer" containerID="02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905" Dec 04 06:30:44 crc kubenswrapper[4685]: E1204 06:30:44.396329 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905\": container with ID starting with 02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905 not found: ID does not exist" containerID="02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.396373 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905"} err="failed to get container status \"02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905\": rpc error: code = NotFound desc = could not find container \"02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905\": container with ID starting with 02bc00a1db50f23e80c09d83d7f9137452b4427dbe0708ada4af2fc586e36905 not found: ID does not exist" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.396404 4685 scope.go:117] "RemoveContainer" containerID="6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af" Dec 04 06:30:44 crc kubenswrapper[4685]: E1204 06:30:44.402125 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af\": container with ID starting with 6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af not found: ID does not exist" containerID="6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.402162 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af"} err="failed to get container status \"6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af\": rpc error: code = NotFound desc = could not find container \"6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af\": container with ID starting with 6db071e50e6eccdeaed8f9d57728a9372bf3eb2765775bbee9b457068bf1c2af not found: ID does not exist" Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.445976 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-2g4gj"] Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.489471 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-tvq7r"] Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.495629 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-tvq7r"] Dec 04 06:30:44 crc kubenswrapper[4685]: I1204 06:30:44.612273 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:44 crc kubenswrapper[4685]: W1204 06:30:44.671571 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd33b8396_6ff8_4bac_90b6_f6f4da823e3b.slice/crio-6879e25a1e7e91d6fc6b242ebcc85d594b7091c8946083d65a599908fb864704 WatchSource:0}: Error finding container 6879e25a1e7e91d6fc6b242ebcc85d594b7091c8946083d65a599908fb864704: Status 404 returned error can't find the container with id 6879e25a1e7e91d6fc6b242ebcc85d594b7091c8946083d65a599908fb864704 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.028923 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b6b84b88b-bjcmm" event={"ID":"66becbc7-a80f-45d5-a711-f0c50c304034","Type":"ContainerStarted","Data":"d7dad80017d54b72ad43dc9e7de96b8117f516e288980d1dcaaf557847a54fe5"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.029208 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.029223 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.034333 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" event={"ID":"1d7a178f-f4ba-4a18-ab97-f2d256e1edc1","Type":"ContainerStarted","Data":"32d13d7442515575d7b54f433d574cd0799cdb54348f85a17fedfdb3d08df5e2"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.047668 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-b6b84b88b-bjcmm" podStartSLOduration=11.04765418 podStartE2EDuration="11.04765418s" podCreationTimestamp="2025-12-04 06:30:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:45.046807413 +0000 UTC m=+1182.235038198" watchObservedRunningTime="2025-12-04 06:30:45.04765418 +0000 UTC m=+1182.235884955" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.048047 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33b8396-6ff8-4bac-90b6-f6f4da823e3b","Type":"ContainerStarted","Data":"6879e25a1e7e91d6fc6b242ebcc85d594b7091c8946083d65a599908fb864704"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.058343 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b690f179-121b-48b5-87bc-16d8c9dfb89e","Type":"ContainerStarted","Data":"259472ccd9669b1816989300873db6d6f4baf5270304a7f491549f901893c153"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.070202 4685 generic.go:334] "Generic (PLEG): container finished" podID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerID="95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0" exitCode=0 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.071186 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" event={"ID":"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d","Type":"ContainerDied","Data":"95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.073427 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" event={"ID":"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d","Type":"ContainerStarted","Data":"c5beb3197900020d80843b999bb22944bff0a80cfd7e7c4ac715dc9f1597cb70"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.074479 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-64bd497db6-z9fft" podStartSLOduration=11.037341624 podStartE2EDuration="14.074462173s" podCreationTimestamp="2025-12-04 06:30:31 +0000 UTC" firstStartedPulling="2025-12-04 06:30:33.203389576 +0000 UTC m=+1170.391620351" lastFinishedPulling="2025-12-04 06:30:36.240510135 +0000 UTC m=+1173.428740900" observedRunningTime="2025-12-04 06:30:45.069515368 +0000 UTC m=+1182.257746153" watchObservedRunningTime="2025-12-04 06:30:45.074462173 +0000 UTC m=+1182.262692948" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.083249 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerStarted","Data":"b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.083486 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-central-agent" containerID="cri-o://56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3" gracePeriod=30 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.083691 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="proxy-httpd" containerID="cri-o://b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1" gracePeriod=30 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.083745 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="sg-core" containerID="cri-o://99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a" gracePeriod=30 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.083779 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-notification-agent" containerID="cri-o://88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a" gracePeriod=30 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.083887 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.110102 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84f9d55fbc-qkcbh" event={"ID":"a1d37923-dce0-47f0-990f-92efc1754323","Type":"ContainerStarted","Data":"5a71a4f739e10eb08e7a50091ffcebba9fdb0610885a91e8b480f131ba99458e"} Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.164882 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.35205297 podStartE2EDuration="1m2.164857568s" podCreationTimestamp="2025-12-04 06:29:43 +0000 UTC" firstStartedPulling="2025-12-04 06:29:44.918860176 +0000 UTC m=+1122.107090951" lastFinishedPulling="2025-12-04 06:30:43.731664774 +0000 UTC m=+1180.919895549" observedRunningTime="2025-12-04 06:30:45.138682785 +0000 UTC m=+1182.326913560" watchObservedRunningTime="2025-12-04 06:30:45.164857568 +0000 UTC m=+1182.353088343" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.205265 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-84f9d55fbc-qkcbh" podStartSLOduration=10.868243464 podStartE2EDuration="14.205242249s" podCreationTimestamp="2025-12-04 06:30:31 +0000 UTC" firstStartedPulling="2025-12-04 06:30:32.893825885 +0000 UTC m=+1170.082056660" lastFinishedPulling="2025-12-04 06:30:36.23082467 +0000 UTC m=+1173.419055445" observedRunningTime="2025-12-04 06:30:45.181990847 +0000 UTC m=+1182.370221622" watchObservedRunningTime="2025-12-04 06:30:45.205242249 +0000 UTC m=+1182.393473024" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.231751 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" path="/var/lib/kubelet/pods/c52e4c01-ae05-4a0b-b245-04e570525017/volumes" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.452845 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-964564b4f-vkm6n" Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.552392 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6985cd56b8-9gnlv"] Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.563819 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6985cd56b8-9gnlv" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-api" containerID="cri-o://7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9" gracePeriod=30 Dec 04 06:30:45 crc kubenswrapper[4685]: I1204 06:30:45.564277 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6985cd56b8-9gnlv" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-httpd" containerID="cri-o://bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770" gracePeriod=30 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.015624 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.199681 4685 generic.go:334] "Generic (PLEG): container finished" podID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerID="b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1" exitCode=0 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.199718 4685 generic.go:334] "Generic (PLEG): container finished" podID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerID="99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a" exitCode=2 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.199726 4685 generic.go:334] "Generic (PLEG): container finished" podID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerID="56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3" exitCode=0 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.199790 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerDied","Data":"b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.199814 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerDied","Data":"99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.199824 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerDied","Data":"56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.212369 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.215679 4685 generic.go:334] "Generic (PLEG): container finished" podID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerID="5fba8a68b2f3c8900cd7c2d81f68ffadb6367ed2359e9e2be619f08f13c5ef30" exitCode=137 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.215706 4685 generic.go:334] "Generic (PLEG): container finished" podID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerID="c78490ec1b2576e8d2d66ed40080d3ef47ad819e6a75fa496653827bd45b1d9c" exitCode=137 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.215745 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddd9d7c47-dwpfz" event={"ID":"b56721cb-e8ce-4106-920b-0b735ef29a38","Type":"ContainerDied","Data":"5fba8a68b2f3c8900cd7c2d81f68ffadb6367ed2359e9e2be619f08f13c5ef30"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.215772 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddd9d7c47-dwpfz" event={"ID":"b56721cb-e8ce-4106-920b-0b735ef29a38","Type":"ContainerDied","Data":"c78490ec1b2576e8d2d66ed40080d3ef47ad819e6a75fa496653827bd45b1d9c"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.220432 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33b8396-6ff8-4bac-90b6-f6f4da823e3b","Type":"ContainerStarted","Data":"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.225989 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" event={"ID":"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d","Type":"ContainerStarted","Data":"365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.226984 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.241486 4685 generic.go:334] "Generic (PLEG): container finished" podID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerID="bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770" exitCode=0 Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.241851 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6985cd56b8-9gnlv" event={"ID":"c657cb75-b17e-41fb-8192-2a6cf3ccbe85","Type":"ContainerDied","Data":"bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770"} Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.249639 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" podStartSLOduration=3.249622179 podStartE2EDuration="3.249622179s" podCreationTimestamp="2025-12-04 06:30:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:46.245072785 +0000 UTC m=+1183.433303560" watchObservedRunningTime="2025-12-04 06:30:46.249622179 +0000 UTC m=+1183.437852954" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.332116 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.398782 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-scripts\") pod \"b56721cb-e8ce-4106-920b-0b735ef29a38\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.398828 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b56721cb-e8ce-4106-920b-0b735ef29a38-horizon-secret-key\") pod \"b56721cb-e8ce-4106-920b-0b735ef29a38\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.398872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2dzb\" (UniqueName: \"kubernetes.io/projected/b56721cb-e8ce-4106-920b-0b735ef29a38-kube-api-access-f2dzb\") pod \"b56721cb-e8ce-4106-920b-0b735ef29a38\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.398891 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b56721cb-e8ce-4106-920b-0b735ef29a38-logs\") pod \"b56721cb-e8ce-4106-920b-0b735ef29a38\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.398949 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-config-data\") pod \"b56721cb-e8ce-4106-920b-0b735ef29a38\" (UID: \"b56721cb-e8ce-4106-920b-0b735ef29a38\") " Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.402064 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.405193 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b56721cb-e8ce-4106-920b-0b735ef29a38-logs" (OuterVolumeSpecName: "logs") pod "b56721cb-e8ce-4106-920b-0b735ef29a38" (UID: "b56721cb-e8ce-4106-920b-0b735ef29a38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.413319 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b56721cb-e8ce-4106-920b-0b735ef29a38-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b56721cb-e8ce-4106-920b-0b735ef29a38" (UID: "b56721cb-e8ce-4106-920b-0b735ef29a38"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.425521 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b56721cb-e8ce-4106-920b-0b735ef29a38-kube-api-access-f2dzb" (OuterVolumeSpecName: "kube-api-access-f2dzb") pod "b56721cb-e8ce-4106-920b-0b735ef29a38" (UID: "b56721cb-e8ce-4106-920b-0b735ef29a38"). InnerVolumeSpecName "kube-api-access-f2dzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.475987 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-config-data" (OuterVolumeSpecName: "config-data") pod "b56721cb-e8ce-4106-920b-0b735ef29a38" (UID: "b56721cb-e8ce-4106-920b-0b735ef29a38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.495080 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.500992 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.501019 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b56721cb-e8ce-4106-920b-0b735ef29a38-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.501036 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2dzb\" (UniqueName: \"kubernetes.io/projected/b56721cb-e8ce-4106-920b-0b735ef29a38-kube-api-access-f2dzb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.501046 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b56721cb-e8ce-4106-920b-0b735ef29a38-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.505734 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-scripts" (OuterVolumeSpecName: "scripts") pod "b56721cb-e8ce-4106-920b-0b735ef29a38" (UID: "b56721cb-e8ce-4106-920b-0b735ef29a38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:46 crc kubenswrapper[4685]: I1204 06:30:46.602734 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b56721cb-e8ce-4106-920b-0b735ef29a38-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.120390 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.255369 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddd9d7c47-dwpfz" event={"ID":"b56721cb-e8ce-4106-920b-0b735ef29a38","Type":"ContainerDied","Data":"99a13dcde48e5e5ae7f7a6500673b9221155b2f40d39f1e55b7ad7bf161fd97c"} Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.255432 4685 scope.go:117] "RemoveContainer" containerID="5fba8a68b2f3c8900cd7c2d81f68ffadb6367ed2359e9e2be619f08f13c5ef30" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.255517 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddd9d7c47-dwpfz" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.285180 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ddd9d7c47-dwpfz"] Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.292313 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6ddd9d7c47-dwpfz"] Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.296680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33b8396-6ff8-4bac-90b6-f6f4da823e3b","Type":"ContainerStarted","Data":"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6"} Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.296824 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api-log" containerID="cri-o://0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d" gracePeriod=30 Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.296927 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.297049 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api" containerID="cri-o://2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6" gracePeriod=30 Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.303128 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b690f179-121b-48b5-87bc-16d8c9dfb89e","Type":"ContainerStarted","Data":"1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816"} Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.508875 4685 scope.go:117] "RemoveContainer" containerID="c78490ec1b2576e8d2d66ed40080d3ef47ad819e6a75fa496653827bd45b1d9c" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.742788 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.782437 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.782385955 podStartE2EDuration="4.782385955s" podCreationTimestamp="2025-12-04 06:30:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:47.331081815 +0000 UTC m=+1184.519312600" watchObservedRunningTime="2025-12-04 06:30:47.782385955 +0000 UTC m=+1184.970616730" Dec 04 06:30:47 crc kubenswrapper[4685]: I1204 06:30:47.984849 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146431 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wtdm\" (UniqueName: \"kubernetes.io/projected/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-kube-api-access-9wtdm\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146474 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146515 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-scripts\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146560 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-logs\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146616 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-combined-ca-bundle\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146722 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-etc-machine-id\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146763 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data-custom\") pod \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\" (UID: \"d33b8396-6ff8-4bac-90b6-f6f4da823e3b\") " Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.146980 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-logs" (OuterVolumeSpecName: "logs") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.147490 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.147532 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.153796 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-scripts" (OuterVolumeSpecName: "scripts") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.155591 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.155670 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-kube-api-access-9wtdm" (OuterVolumeSpecName: "kube-api-access-9wtdm") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "kube-api-access-9wtdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.178537 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.231278 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data" (OuterVolumeSpecName: "config-data") pod "d33b8396-6ff8-4bac-90b6-f6f4da823e3b" (UID: "d33b8396-6ff8-4bac-90b6-f6f4da823e3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.252606 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.252642 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wtdm\" (UniqueName: \"kubernetes.io/projected/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-kube-api-access-9wtdm\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.252652 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.252661 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.252670 4685 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.252678 4685 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33b8396-6ff8-4bac-90b6-f6f4da823e3b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.313008 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b690f179-121b-48b5-87bc-16d8c9dfb89e","Type":"ContainerStarted","Data":"aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462"} Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317067 4685 generic.go:334] "Generic (PLEG): container finished" podID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerID="2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6" exitCode=0 Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317099 4685 generic.go:334] "Generic (PLEG): container finished" podID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerID="0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d" exitCode=143 Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317138 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317219 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33b8396-6ff8-4bac-90b6-f6f4da823e3b","Type":"ContainerDied","Data":"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6"} Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317274 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33b8396-6ff8-4bac-90b6-f6f4da823e3b","Type":"ContainerDied","Data":"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d"} Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317289 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33b8396-6ff8-4bac-90b6-f6f4da823e3b","Type":"ContainerDied","Data":"6879e25a1e7e91d6fc6b242ebcc85d594b7091c8946083d65a599908fb864704"} Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.317311 4685 scope.go:117] "RemoveContainer" containerID="2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.357983 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.861400648 podStartE2EDuration="5.357967225s" podCreationTimestamp="2025-12-04 06:30:43 +0000 UTC" firstStartedPulling="2025-12-04 06:30:44.401479609 +0000 UTC m=+1181.589710384" lastFinishedPulling="2025-12-04 06:30:45.898046186 +0000 UTC m=+1183.086276961" observedRunningTime="2025-12-04 06:30:48.345261455 +0000 UTC m=+1185.533492230" watchObservedRunningTime="2025-12-04 06:30:48.357967225 +0000 UTC m=+1185.546198000" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.369617 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.376754 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.385367 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.388464 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api-log" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388487 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api-log" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.388505 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" containerName="init" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388511 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" containerName="init" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.388523 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388529 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.388540 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon-log" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388546 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon-log" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.388558 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388564 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.388578 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" containerName="dnsmasq-dns" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388586 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" containerName="dnsmasq-dns" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388766 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon-log" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388783 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388794 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c52e4c01-ae05-4a0b-b245-04e570525017" containerName="dnsmasq-dns" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388806 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" containerName="cinder-api-log" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.388822 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" containerName="horizon" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.389751 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.391724 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.391886 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.393270 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.398518 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.456996 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad866e82-c1d3-4a38-953e-917ef323430b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457109 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-config-data\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457153 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457175 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqchj\" (UniqueName: \"kubernetes.io/projected/ad866e82-c1d3-4a38-953e-917ef323430b-kube-api-access-fqchj\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457455 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457597 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-scripts\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457668 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-config-data-custom\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457715 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.457825 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad866e82-c1d3-4a38-953e-917ef323430b-logs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.470951 4685 scope.go:117] "RemoveContainer" containerID="0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.494686 4685 scope.go:117] "RemoveContainer" containerID="2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.495075 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6\": container with ID starting with 2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6 not found: ID does not exist" containerID="2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.495116 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6"} err="failed to get container status \"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6\": rpc error: code = NotFound desc = could not find container \"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6\": container with ID starting with 2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6 not found: ID does not exist" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.495144 4685 scope.go:117] "RemoveContainer" containerID="0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d" Dec 04 06:30:48 crc kubenswrapper[4685]: E1204 06:30:48.495722 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d\": container with ID starting with 0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d not found: ID does not exist" containerID="0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.495768 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d"} err="failed to get container status \"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d\": rpc error: code = NotFound desc = could not find container \"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d\": container with ID starting with 0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d not found: ID does not exist" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.495796 4685 scope.go:117] "RemoveContainer" containerID="2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.496210 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6"} err="failed to get container status \"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6\": rpc error: code = NotFound desc = could not find container \"2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6\": container with ID starting with 2f0824df9056f8c4e83877edf292714c6388a968450ca05ea868ab99222e67f6 not found: ID does not exist" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.496243 4685 scope.go:117] "RemoveContainer" containerID="0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.496487 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d"} err="failed to get container status \"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d\": rpc error: code = NotFound desc = could not find container \"0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d\": container with ID starting with 0101da79610237b50718dd2795d811da3d81846a7677090fe6d76f4cbe7d6d3d not found: ID does not exist" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558471 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-scripts\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558514 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-config-data-custom\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558538 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558574 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad866e82-c1d3-4a38-953e-917ef323430b-logs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558598 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad866e82-c1d3-4a38-953e-917ef323430b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558634 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-config-data\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558653 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558671 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqchj\" (UniqueName: \"kubernetes.io/projected/ad866e82-c1d3-4a38-953e-917ef323430b-kube-api-access-fqchj\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.558723 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.559107 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad866e82-c1d3-4a38-953e-917ef323430b-logs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.559446 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad866e82-c1d3-4a38-953e-917ef323430b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.565398 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-scripts\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.565971 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-config-data-custom\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.566849 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.567944 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-config-data\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.571359 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.572768 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad866e82-c1d3-4a38-953e-917ef323430b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.586862 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqchj\" (UniqueName: \"kubernetes.io/projected/ad866e82-c1d3-4a38-953e-917ef323430b-kube-api-access-fqchj\") pod \"cinder-api-0\" (UID: \"ad866e82-c1d3-4a38-953e-917ef323430b\") " pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.624401 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.724950 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 06:30:48 crc kubenswrapper[4685]: I1204 06:30:48.792369 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.138660 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b56721cb-e8ce-4106-920b-0b735ef29a38" path="/var/lib/kubelet/pods/b56721cb-e8ce-4106-920b-0b735ef29a38/volumes" Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.140139 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d33b8396-6ff8-4bac-90b6-f6f4da823e3b" path="/var/lib/kubelet/pods/d33b8396-6ff8-4bac-90b6-f6f4da823e3b/volumes" Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.209158 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.329371 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ad866e82-c1d3-4a38-953e-917ef323430b","Type":"ContainerStarted","Data":"728fc802b41c173bfd176d78963d1f8c02efde3b39dba765145a991901af5817"} Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.575987 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5c9548d9dd-h5vqf" Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.631668 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bdfd7d6f6-8slln"] Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.631905 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bdfd7d6f6-8slln" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon-log" containerID="cri-o://252f6828706c15e58fcb7a3b491ecee0a37c85fd571fbefe50fcd5492a05e612" gracePeriod=30 Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.632196 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bdfd7d6f6-8slln" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" containerID="cri-o://7195b2be6943bd0983daf635bd5ec023782817786ac253eb69ce2b406968b974" gracePeriod=30 Dec 04 06:30:49 crc kubenswrapper[4685]: I1204 06:30:49.919303 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.089687 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-run-httpd\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.089890 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-combined-ca-bundle\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.089949 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-scripts\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.089989 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-config-data\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.090020 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86lzg\" (UniqueName: \"kubernetes.io/projected/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-kube-api-access-86lzg\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.090070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-log-httpd\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.090149 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-sg-core-conf-yaml\") pod \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\" (UID: \"f9be4721-c5dd-4563-8fe6-9c5bbd92924a\") " Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.090290 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.090699 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.091608 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.095893 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-kube-api-access-86lzg" (OuterVolumeSpecName: "kube-api-access-86lzg") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "kube-api-access-86lzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.098353 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-scripts" (OuterVolumeSpecName: "scripts") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.137002 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.192393 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.192442 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86lzg\" (UniqueName: \"kubernetes.io/projected/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-kube-api-access-86lzg\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.192454 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.192465 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.202741 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.219710 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-config-data" (OuterVolumeSpecName: "config-data") pod "f9be4721-c5dd-4563-8fe6-9c5bbd92924a" (UID: "f9be4721-c5dd-4563-8fe6-9c5bbd92924a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.293560 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.293589 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9be4721-c5dd-4563-8fe6-9c5bbd92924a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.353359 4685 generic.go:334] "Generic (PLEG): container finished" podID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerID="88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a" exitCode=0 Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.353450 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.353454 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerDied","Data":"88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a"} Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.353901 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9be4721-c5dd-4563-8fe6-9c5bbd92924a","Type":"ContainerDied","Data":"20652d1091ddaeab7d547e06db60c052121ee0cea8c3d1ab5e379a90eadc5b13"} Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.353934 4685 scope.go:117] "RemoveContainer" containerID="b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.357951 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ad866e82-c1d3-4a38-953e-917ef323430b","Type":"ContainerStarted","Data":"f83ad4f2a140bf10179272a8d110735b60ba7115a9d2d564f2b8e1583a7b3353"} Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.384720 4685 scope.go:117] "RemoveContainer" containerID="99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.404666 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.467428 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.484844 4685 scope.go:117] "RemoveContainer" containerID="88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.485861 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.486278 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="proxy-httpd" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486299 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="proxy-httpd" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.486320 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-notification-agent" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486327 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-notification-agent" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.486343 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-central-agent" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486349 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-central-agent" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.486358 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="sg-core" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486364 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="sg-core" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486568 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-central-agent" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486586 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="sg-core" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486601 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="proxy-httpd" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.486609 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" containerName="ceilometer-notification-agent" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.488179 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.491216 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.502877 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.508189 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.525294 4685 scope.go:117] "RemoveContainer" containerID="56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.546134 4685 scope.go:117] "RemoveContainer" containerID="b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.546789 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1\": container with ID starting with b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1 not found: ID does not exist" containerID="b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.546839 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1"} err="failed to get container status \"b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1\": rpc error: code = NotFound desc = could not find container \"b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1\": container with ID starting with b42e6479d99e80a23af80977e3e5ab73f91fa8b034c8980e8b1676c99cfb60a1 not found: ID does not exist" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.546861 4685 scope.go:117] "RemoveContainer" containerID="99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.547171 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a\": container with ID starting with 99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a not found: ID does not exist" containerID="99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.547209 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a"} err="failed to get container status \"99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a\": rpc error: code = NotFound desc = could not find container \"99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a\": container with ID starting with 99a385dcfaffd1030878a6ebcf56998c5aae95ab559a3c3fdddd5b7430fb6b7a not found: ID does not exist" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.547236 4685 scope.go:117] "RemoveContainer" containerID="88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.547655 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a\": container with ID starting with 88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a not found: ID does not exist" containerID="88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.547676 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a"} err="failed to get container status \"88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a\": rpc error: code = NotFound desc = could not find container \"88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a\": container with ID starting with 88f969b651e5ddd6f7cec8e69aaf5de4071aa08dbed135ec17d84249c93d774a not found: ID does not exist" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.547688 4685 scope.go:117] "RemoveContainer" containerID="56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3" Dec 04 06:30:50 crc kubenswrapper[4685]: E1204 06:30:50.547899 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3\": container with ID starting with 56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3 not found: ID does not exist" containerID="56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.547922 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3"} err="failed to get container status \"56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3\": rpc error: code = NotFound desc = could not find container \"56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3\": container with ID starting with 56b2f9160de3488ab3c45291fc733c0b733e361884df0dff42d73a9c90086ec3 not found: ID does not exist" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.599913 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.599999 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-run-httpd\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.601056 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-scripts\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.601098 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-config-data\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.601140 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.601207 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fgdj\" (UniqueName: \"kubernetes.io/projected/58d511b5-5f39-4ea8-84f9-a29f1587368e-kube-api-access-5fgdj\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.601257 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-log-httpd\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703167 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703249 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-run-httpd\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-scripts\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703323 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-config-data\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703353 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703400 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fgdj\" (UniqueName: \"kubernetes.io/projected/58d511b5-5f39-4ea8-84f9-a29f1587368e-kube-api-access-5fgdj\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703461 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-log-httpd\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703836 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-run-httpd\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.703927 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-log-httpd\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.708924 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.709562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.711164 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-config-data\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.712050 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-scripts\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.728050 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fgdj\" (UniqueName: \"kubernetes.io/projected/58d511b5-5f39-4ea8-84f9-a29f1587368e-kube-api-access-5fgdj\") pod \"ceilometer-0\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " pod="openstack/ceilometer-0" Dec 04 06:30:50 crc kubenswrapper[4685]: I1204 06:30:50.819946 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.135894 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9be4721-c5dd-4563-8fe6-9c5bbd92924a" path="/var/lib/kubelet/pods/f9be4721-c5dd-4563-8fe6-9c5bbd92924a/volumes" Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.323128 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.368771 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerStarted","Data":"b12aa6ba60dd1eb9e639d669ef150feb18186c25dfd352c25f2475064bfc05ec"} Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.372770 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ad866e82-c1d3-4a38-953e-917ef323430b","Type":"ContainerStarted","Data":"ff63a422b3ef9e5d0ced70d978918d6992287f7e70b2448e75a52f162a5dbf7c"} Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.373451 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.398901 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.398882573 podStartE2EDuration="3.398882573s" podCreationTimestamp="2025-12-04 06:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:30:51.396677594 +0000 UTC m=+1188.584908389" watchObservedRunningTime="2025-12-04 06:30:51.398882573 +0000 UTC m=+1188.587113348" Dec 04 06:30:51 crc kubenswrapper[4685]: E1204 06:30:51.657098 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc657cb75_b17e_41fb_8192_2a6cf3ccbe85.slice/crio-conmon-7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.972677 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b6b84b88b-bjcmm" Dec 04 06:30:51 crc kubenswrapper[4685]: I1204 06:30:51.992303 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.063792 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c7d4bcdc6-psn87"] Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.064079 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c7d4bcdc6-psn87" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api-log" containerID="cri-o://e14dd9c4bcc207cf304c068d5f2b25174cc34ed3a0111a84d8dd8a22ecac6799" gracePeriod=30 Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.064130 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c7d4bcdc6-psn87" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api" containerID="cri-o://e5811deb4eb8030ae26f6904e5f133b7306db0b535f20eec60092c49123ba493" gracePeriod=30 Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.126897 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzzp6\" (UniqueName: \"kubernetes.io/projected/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-kube-api-access-wzzp6\") pod \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.127033 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-combined-ca-bundle\") pod \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.127092 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-httpd-config\") pod \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.127145 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-config\") pod \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.127210 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-ovndb-tls-certs\") pod \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\" (UID: \"c657cb75-b17e-41fb-8192-2a6cf3ccbe85\") " Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.137559 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c657cb75-b17e-41fb-8192-2a6cf3ccbe85" (UID: "c657cb75-b17e-41fb-8192-2a6cf3ccbe85"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.137693 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-kube-api-access-wzzp6" (OuterVolumeSpecName: "kube-api-access-wzzp6") pod "c657cb75-b17e-41fb-8192-2a6cf3ccbe85" (UID: "c657cb75-b17e-41fb-8192-2a6cf3ccbe85"). InnerVolumeSpecName "kube-api-access-wzzp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.227101 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-config" (OuterVolumeSpecName: "config") pod "c657cb75-b17e-41fb-8192-2a6cf3ccbe85" (UID: "c657cb75-b17e-41fb-8192-2a6cf3ccbe85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.229443 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.229475 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.229487 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzzp6\" (UniqueName: \"kubernetes.io/projected/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-kube-api-access-wzzp6\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.256299 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c657cb75-b17e-41fb-8192-2a6cf3ccbe85" (UID: "c657cb75-b17e-41fb-8192-2a6cf3ccbe85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.323553 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c657cb75-b17e-41fb-8192-2a6cf3ccbe85" (UID: "c657cb75-b17e-41fb-8192-2a6cf3ccbe85"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.331216 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.331249 4685 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c657cb75-b17e-41fb-8192-2a6cf3ccbe85-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.383633 4685 generic.go:334] "Generic (PLEG): container finished" podID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerID="e14dd9c4bcc207cf304c068d5f2b25174cc34ed3a0111a84d8dd8a22ecac6799" exitCode=143 Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.383702 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7d4bcdc6-psn87" event={"ID":"9d68936d-d2fb-489f-91e3-de59fce181a2","Type":"ContainerDied","Data":"e14dd9c4bcc207cf304c068d5f2b25174cc34ed3a0111a84d8dd8a22ecac6799"} Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.385720 4685 generic.go:334] "Generic (PLEG): container finished" podID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerID="7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9" exitCode=0 Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.385766 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6985cd56b8-9gnlv" event={"ID":"c657cb75-b17e-41fb-8192-2a6cf3ccbe85","Type":"ContainerDied","Data":"7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9"} Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.385815 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6985cd56b8-9gnlv" event={"ID":"c657cb75-b17e-41fb-8192-2a6cf3ccbe85","Type":"ContainerDied","Data":"e2d727cf7329731ea689cc3001d03f5971833f132743fc5beb636e3aca5e23aa"} Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.385833 4685 scope.go:117] "RemoveContainer" containerID="bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.385784 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6985cd56b8-9gnlv" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.391120 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerStarted","Data":"0ed325be65dd9582f5d7cceab1ee66bb9dcf354c57f529af3b0bbae0778e58f9"} Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.409152 4685 scope.go:117] "RemoveContainer" containerID="7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.423389 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6985cd56b8-9gnlv"] Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.431834 4685 scope.go:117] "RemoveContainer" containerID="bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770" Dec 04 06:30:52 crc kubenswrapper[4685]: E1204 06:30:52.432965 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770\": container with ID starting with bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770 not found: ID does not exist" containerID="bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.433034 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770"} err="failed to get container status \"bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770\": rpc error: code = NotFound desc = could not find container \"bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770\": container with ID starting with bd40d5c2858d2f95a3bd4351e6c34da6f34779b9222d9ae996610c8b3d7a5770 not found: ID does not exist" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.433069 4685 scope.go:117] "RemoveContainer" containerID="7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9" Dec 04 06:30:52 crc kubenswrapper[4685]: E1204 06:30:52.438252 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9\": container with ID starting with 7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9 not found: ID does not exist" containerID="7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.438390 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9"} err="failed to get container status \"7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9\": rpc error: code = NotFound desc = could not find container \"7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9\": container with ID starting with 7edb10b1d331ed2a257f0ddaf54784426263cd471cb9c63d342be44dc56f29c9 not found: ID does not exist" Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.448876 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6985cd56b8-9gnlv"] Dec 04 06:30:52 crc kubenswrapper[4685]: I1204 06:30:52.903627 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6bdfd7d6f6-8slln" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.139565 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" path="/var/lib/kubelet/pods/c657cb75-b17e-41fb-8192-2a6cf3ccbe85/volumes" Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.401315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerStarted","Data":"dafb7bf39e28a94dddd392c969e530dae6832d12dc474bb4b856873dcabd113e"} Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.403081 4685 generic.go:334] "Generic (PLEG): container finished" podID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerID="7195b2be6943bd0983daf635bd5ec023782817786ac253eb69ce2b406968b974" exitCode=0 Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.403109 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bdfd7d6f6-8slln" event={"ID":"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8","Type":"ContainerDied","Data":"7195b2be6943bd0983daf635bd5ec023782817786ac253eb69ce2b406968b974"} Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.844744 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.909199 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-ssdgf"] Dec 04 06:30:53 crc kubenswrapper[4685]: I1204 06:30:53.909430 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerName="dnsmasq-dns" containerID="cri-o://3e3bd003cb39deea54c92c9d70e098dacbe512bcc173de0ea794de690f77dea4" gracePeriod=10 Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.111844 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.164927 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.411872 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerStarted","Data":"701540091a402585044499802211f8b4eb9b948744e08d903570780edbd2eebc"} Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.414346 4685 generic.go:334] "Generic (PLEG): container finished" podID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerID="3e3bd003cb39deea54c92c9d70e098dacbe512bcc173de0ea794de690f77dea4" exitCode=0 Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.414609 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="cinder-scheduler" containerID="cri-o://1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816" gracePeriod=30 Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.414683 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" event={"ID":"17281897-9381-4ab1-a79f-9cd39a8b5508","Type":"ContainerDied","Data":"3e3bd003cb39deea54c92c9d70e098dacbe512bcc173de0ea794de690f77dea4"} Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.415037 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="probe" containerID="cri-o://aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462" gracePeriod=30 Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.673463 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.795484 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-svc\") pod \"17281897-9381-4ab1-a79f-9cd39a8b5508\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.795554 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-sb\") pod \"17281897-9381-4ab1-a79f-9cd39a8b5508\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.795601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrhgm\" (UniqueName: \"kubernetes.io/projected/17281897-9381-4ab1-a79f-9cd39a8b5508-kube-api-access-vrhgm\") pod \"17281897-9381-4ab1-a79f-9cd39a8b5508\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.795645 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-swift-storage-0\") pod \"17281897-9381-4ab1-a79f-9cd39a8b5508\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.795670 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-nb\") pod \"17281897-9381-4ab1-a79f-9cd39a8b5508\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.795824 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-config\") pod \"17281897-9381-4ab1-a79f-9cd39a8b5508\" (UID: \"17281897-9381-4ab1-a79f-9cd39a8b5508\") " Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.801563 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17281897-9381-4ab1-a79f-9cd39a8b5508-kube-api-access-vrhgm" (OuterVolumeSpecName: "kube-api-access-vrhgm") pod "17281897-9381-4ab1-a79f-9cd39a8b5508" (UID: "17281897-9381-4ab1-a79f-9cd39a8b5508"). InnerVolumeSpecName "kube-api-access-vrhgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.883042 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-config" (OuterVolumeSpecName: "config") pod "17281897-9381-4ab1-a79f-9cd39a8b5508" (UID: "17281897-9381-4ab1-a79f-9cd39a8b5508"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.893354 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17281897-9381-4ab1-a79f-9cd39a8b5508" (UID: "17281897-9381-4ab1-a79f-9cd39a8b5508"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.897884 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17281897-9381-4ab1-a79f-9cd39a8b5508" (UID: "17281897-9381-4ab1-a79f-9cd39a8b5508"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.899140 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.899171 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.899181 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrhgm\" (UniqueName: \"kubernetes.io/projected/17281897-9381-4ab1-a79f-9cd39a8b5508-kube-api-access-vrhgm\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.899191 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.911371 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17281897-9381-4ab1-a79f-9cd39a8b5508" (UID: "17281897-9381-4ab1-a79f-9cd39a8b5508"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:54 crc kubenswrapper[4685]: I1204 06:30:54.914621 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17281897-9381-4ab1-a79f-9cd39a8b5508" (UID: "17281897-9381-4ab1-a79f-9cd39a8b5508"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.001041 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.001074 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17281897-9381-4ab1-a79f-9cd39a8b5508-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.177976 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c7d4bcdc6-psn87" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:45290->10.217.0.160:9311: read: connection reset by peer" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.178222 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c7d4bcdc6-psn87" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:45296->10.217.0.160:9311: read: connection reset by peer" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.435686 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.455611 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerStarted","Data":"1941db914492dacfb073d1f6d709ad5371ccbb812f95c6e627e286ac180d8f67"} Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.455947 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.475009 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" event={"ID":"17281897-9381-4ab1-a79f-9cd39a8b5508","Type":"ContainerDied","Data":"51580cfdd26d0ca0fa7e8e5dc00d9c650bbc187e1febbba176a008b448385186"} Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.475062 4685 scope.go:117] "RemoveContainer" containerID="3e3bd003cb39deea54c92c9d70e098dacbe512bcc173de0ea794de690f77dea4" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.475172 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-ssdgf" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.523973 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6bc7689ffb-9thrj" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.532112 4685 generic.go:334] "Generic (PLEG): container finished" podID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerID="e5811deb4eb8030ae26f6904e5f133b7306db0b535f20eec60092c49123ba493" exitCode=0 Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.532228 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7d4bcdc6-psn87" event={"ID":"9d68936d-d2fb-489f-91e3-de59fce181a2","Type":"ContainerDied","Data":"e5811deb4eb8030ae26f6904e5f133b7306db0b535f20eec60092c49123ba493"} Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.541099 4685 generic.go:334] "Generic (PLEG): container finished" podID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerID="aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462" exitCode=0 Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.541148 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b690f179-121b-48b5-87bc-16d8c9dfb89e","Type":"ContainerDied","Data":"aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462"} Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.581135 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.407679624 podStartE2EDuration="5.581109781s" podCreationTimestamp="2025-12-04 06:30:50 +0000 UTC" firstStartedPulling="2025-12-04 06:30:51.333066203 +0000 UTC m=+1188.521296968" lastFinishedPulling="2025-12-04 06:30:54.50649635 +0000 UTC m=+1191.694727125" observedRunningTime="2025-12-04 06:30:55.503146798 +0000 UTC m=+1192.691377573" watchObservedRunningTime="2025-12-04 06:30:55.581109781 +0000 UTC m=+1192.769340556" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.596496 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-ssdgf"] Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.603316 4685 scope.go:117] "RemoveContainer" containerID="296cb7cbca21efad93ddd8a273033cd984e544b2321dadb75a686d630659a213" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.614243 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-ssdgf"] Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.715206 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.825807 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d68936d-d2fb-489f-91e3-de59fce181a2-logs\") pod \"9d68936d-d2fb-489f-91e3-de59fce181a2\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.825953 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data-custom\") pod \"9d68936d-d2fb-489f-91e3-de59fce181a2\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.825990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tbz2\" (UniqueName: \"kubernetes.io/projected/9d68936d-d2fb-489f-91e3-de59fce181a2-kube-api-access-8tbz2\") pod \"9d68936d-d2fb-489f-91e3-de59fce181a2\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.826038 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data\") pod \"9d68936d-d2fb-489f-91e3-de59fce181a2\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.826069 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-combined-ca-bundle\") pod \"9d68936d-d2fb-489f-91e3-de59fce181a2\" (UID: \"9d68936d-d2fb-489f-91e3-de59fce181a2\") " Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.826446 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d68936d-d2fb-489f-91e3-de59fce181a2-logs" (OuterVolumeSpecName: "logs") pod "9d68936d-d2fb-489f-91e3-de59fce181a2" (UID: "9d68936d-d2fb-489f-91e3-de59fce181a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.831879 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d68936d-d2fb-489f-91e3-de59fce181a2-kube-api-access-8tbz2" (OuterVolumeSpecName: "kube-api-access-8tbz2") pod "9d68936d-d2fb-489f-91e3-de59fce181a2" (UID: "9d68936d-d2fb-489f-91e3-de59fce181a2"). InnerVolumeSpecName "kube-api-access-8tbz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.832615 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9d68936d-d2fb-489f-91e3-de59fce181a2" (UID: "9d68936d-d2fb-489f-91e3-de59fce181a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.846693 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-679cdf976b-2nlz9" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.858644 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d68936d-d2fb-489f-91e3-de59fce181a2" (UID: "9d68936d-d2fb-489f-91e3-de59fce181a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.892935 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data" (OuterVolumeSpecName: "config-data") pod "9d68936d-d2fb-489f-91e3-de59fce181a2" (UID: "9d68936d-d2fb-489f-91e3-de59fce181a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.929040 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d68936d-d2fb-489f-91e3-de59fce181a2-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.929085 4685 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.929117 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tbz2\" (UniqueName: \"kubernetes.io/projected/9d68936d-d2fb-489f-91e3-de59fce181a2-kube-api-access-8tbz2\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.929126 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:55 crc kubenswrapper[4685]: I1204 06:30:55.929136 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d68936d-d2fb-489f-91e3-de59fce181a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:56 crc kubenswrapper[4685]: I1204 06:30:56.564542 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7d4bcdc6-psn87" event={"ID":"9d68936d-d2fb-489f-91e3-de59fce181a2","Type":"ContainerDied","Data":"79e29e3e7a5584aca2d3e0f8c3e1c7b191194abe8456e1722680833d9ff2fb65"} Dec 04 06:30:56 crc kubenswrapper[4685]: I1204 06:30:56.564625 4685 scope.go:117] "RemoveContainer" containerID="e5811deb4eb8030ae26f6904e5f133b7306db0b535f20eec60092c49123ba493" Dec 04 06:30:56 crc kubenswrapper[4685]: I1204 06:30:56.564825 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7d4bcdc6-psn87" Dec 04 06:30:56 crc kubenswrapper[4685]: I1204 06:30:56.612620 4685 scope.go:117] "RemoveContainer" containerID="e14dd9c4bcc207cf304c068d5f2b25174cc34ed3a0111a84d8dd8a22ecac6799" Dec 04 06:30:56 crc kubenswrapper[4685]: I1204 06:30:56.638190 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c7d4bcdc6-psn87"] Dec 04 06:30:56 crc kubenswrapper[4685]: I1204 06:30:56.646825 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c7d4bcdc6-psn87"] Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.136953 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" path="/var/lib/kubelet/pods/17281897-9381-4ab1-a79f-9cd39a8b5508/volumes" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.138103 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" path="/var/lib/kubelet/pods/9d68936d-d2fb-489f-91e3-de59fce181a2/volumes" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.956397 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 06:30:57 crc kubenswrapper[4685]: E1204 06:30:57.957049 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957116 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api" Dec 04 06:30:57 crc kubenswrapper[4685]: E1204 06:30:57.957185 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-api" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957276 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-api" Dec 04 06:30:57 crc kubenswrapper[4685]: E1204 06:30:57.957337 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerName="dnsmasq-dns" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957393 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerName="dnsmasq-dns" Dec 04 06:30:57 crc kubenswrapper[4685]: E1204 06:30:57.957471 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-httpd" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957521 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-httpd" Dec 04 06:30:57 crc kubenswrapper[4685]: E1204 06:30:57.957576 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api-log" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957652 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api-log" Dec 04 06:30:57 crc kubenswrapper[4685]: E1204 06:30:57.957717 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerName="init" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957767 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerName="init" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.957971 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api-log" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.958048 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-api" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.958104 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c657cb75-b17e-41fb-8192-2a6cf3ccbe85" containerName="neutron-httpd" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.958167 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="17281897-9381-4ab1-a79f-9cd39a8b5508" containerName="dnsmasq-dns" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.958232 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d68936d-d2fb-489f-91e3-de59fce181a2" containerName="barbican-api" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.958878 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.965348 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.965470 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.965627 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bqrtj" Dec 04 06:30:57 crc kubenswrapper[4685]: I1204 06:30:57.969568 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.069810 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4a5bce-68e1-448c-a00c-d7fc97e69566-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.070105 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a4a5bce-68e1-448c-a00c-d7fc97e69566-openstack-config\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.070137 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvvvv\" (UniqueName: \"kubernetes.io/projected/6a4a5bce-68e1-448c-a00c-d7fc97e69566-kube-api-access-fvvvv\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.070174 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a4a5bce-68e1-448c-a00c-d7fc97e69566-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.175772 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4a5bce-68e1-448c-a00c-d7fc97e69566-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.175826 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a4a5bce-68e1-448c-a00c-d7fc97e69566-openstack-config\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.175847 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvvvv\" (UniqueName: \"kubernetes.io/projected/6a4a5bce-68e1-448c-a00c-d7fc97e69566-kube-api-access-fvvvv\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.175870 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a4a5bce-68e1-448c-a00c-d7fc97e69566-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.177577 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a4a5bce-68e1-448c-a00c-d7fc97e69566-openstack-config\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.185781 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a4a5bce-68e1-448c-a00c-d7fc97e69566-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.200003 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4a5bce-68e1-448c-a00c-d7fc97e69566-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.201024 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvvvv\" (UniqueName: \"kubernetes.io/projected/6a4a5bce-68e1-448c-a00c-d7fc97e69566-kube-api-access-fvvvv\") pod \"openstackclient\" (UID: \"6a4a5bce-68e1-448c-a00c-d7fc97e69566\") " pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.287332 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.414868 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.484813 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b690f179-121b-48b5-87bc-16d8c9dfb89e-etc-machine-id\") pod \"b690f179-121b-48b5-87bc-16d8c9dfb89e\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.484936 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b690f179-121b-48b5-87bc-16d8c9dfb89e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b690f179-121b-48b5-87bc-16d8c9dfb89e" (UID: "b690f179-121b-48b5-87bc-16d8c9dfb89e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.484990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfksm\" (UniqueName: \"kubernetes.io/projected/b690f179-121b-48b5-87bc-16d8c9dfb89e-kube-api-access-xfksm\") pod \"b690f179-121b-48b5-87bc-16d8c9dfb89e\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.485029 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-combined-ca-bundle\") pod \"b690f179-121b-48b5-87bc-16d8c9dfb89e\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.485081 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data-custom\") pod \"b690f179-121b-48b5-87bc-16d8c9dfb89e\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.485135 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-scripts\") pod \"b690f179-121b-48b5-87bc-16d8c9dfb89e\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.485188 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data\") pod \"b690f179-121b-48b5-87bc-16d8c9dfb89e\" (UID: \"b690f179-121b-48b5-87bc-16d8c9dfb89e\") " Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.485591 4685 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b690f179-121b-48b5-87bc-16d8c9dfb89e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.500652 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b690f179-121b-48b5-87bc-16d8c9dfb89e" (UID: "b690f179-121b-48b5-87bc-16d8c9dfb89e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.500794 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b690f179-121b-48b5-87bc-16d8c9dfb89e-kube-api-access-xfksm" (OuterVolumeSpecName: "kube-api-access-xfksm") pod "b690f179-121b-48b5-87bc-16d8c9dfb89e" (UID: "b690f179-121b-48b5-87bc-16d8c9dfb89e"). InnerVolumeSpecName "kube-api-access-xfksm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.507550 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-scripts" (OuterVolumeSpecName: "scripts") pod "b690f179-121b-48b5-87bc-16d8c9dfb89e" (UID: "b690f179-121b-48b5-87bc-16d8c9dfb89e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.550718 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b690f179-121b-48b5-87bc-16d8c9dfb89e" (UID: "b690f179-121b-48b5-87bc-16d8c9dfb89e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.591813 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfksm\" (UniqueName: \"kubernetes.io/projected/b690f179-121b-48b5-87bc-16d8c9dfb89e-kube-api-access-xfksm\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.591849 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.591859 4685 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.591870 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.598276 4685 generic.go:334] "Generic (PLEG): container finished" podID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerID="1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816" exitCode=0 Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.598317 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b690f179-121b-48b5-87bc-16d8c9dfb89e","Type":"ContainerDied","Data":"1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816"} Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.598343 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b690f179-121b-48b5-87bc-16d8c9dfb89e","Type":"ContainerDied","Data":"259472ccd9669b1816989300873db6d6f4baf5270304a7f491549f901893c153"} Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.598340 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.598373 4685 scope.go:117] "RemoveContainer" containerID="aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.603461 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data" (OuterVolumeSpecName: "config-data") pod "b690f179-121b-48b5-87bc-16d8c9dfb89e" (UID: "b690f179-121b-48b5-87bc-16d8c9dfb89e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.627378 4685 scope.go:117] "RemoveContainer" containerID="1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.652291 4685 scope.go:117] "RemoveContainer" containerID="aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462" Dec 04 06:30:58 crc kubenswrapper[4685]: E1204 06:30:58.652824 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462\": container with ID starting with aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462 not found: ID does not exist" containerID="aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.652857 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462"} err="failed to get container status \"aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462\": rpc error: code = NotFound desc = could not find container \"aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462\": container with ID starting with aaadb8fa2239c4447b2418fcc8eca6add21b391a146b675c650d3f2eb9707462 not found: ID does not exist" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.652879 4685 scope.go:117] "RemoveContainer" containerID="1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816" Dec 04 06:30:58 crc kubenswrapper[4685]: E1204 06:30:58.653106 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816\": container with ID starting with 1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816 not found: ID does not exist" containerID="1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.653150 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816"} err="failed to get container status \"1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816\": rpc error: code = NotFound desc = could not find container \"1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816\": container with ID starting with 1466b05afa510ae44e2140556592f24728be244ac3ee3ada19dd1b138e563816 not found: ID does not exist" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.695983 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b690f179-121b-48b5-87bc-16d8c9dfb89e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.792534 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 06:30:58 crc kubenswrapper[4685]: W1204 06:30:58.796276 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a4a5bce_68e1_448c_a00c_d7fc97e69566.slice/crio-4a2e92c7de645d020eac4237c63dcb33e4cc1d2fc86cd2ead4143a9b577a15a2 WatchSource:0}: Error finding container 4a2e92c7de645d020eac4237c63dcb33e4cc1d2fc86cd2ead4143a9b577a15a2: Status 404 returned error can't find the container with id 4a2e92c7de645d020eac4237c63dcb33e4cc1d2fc86cd2ead4143a9b577a15a2 Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.928638 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.943090 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.962129 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:58 crc kubenswrapper[4685]: E1204 06:30:58.962953 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="cinder-scheduler" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.962976 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="cinder-scheduler" Dec 04 06:30:58 crc kubenswrapper[4685]: E1204 06:30:58.963010 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="probe" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.963020 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="probe" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.963476 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="probe" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.963512 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" containerName="cinder-scheduler" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.967651 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.975535 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 06:30:58 crc kubenswrapper[4685]: I1204 06:30:58.989053 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.101608 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.101669 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67076e23-eac4-4e2a-944a-8687060343ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.101689 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.101932 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.102047 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.102087 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4wwx\" (UniqueName: \"kubernetes.io/projected/67076e23-eac4-4e2a-944a-8687060343ad-kube-api-access-c4wwx\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.139195 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b690f179-121b-48b5-87bc-16d8c9dfb89e" path="/var/lib/kubelet/pods/b690f179-121b-48b5-87bc-16d8c9dfb89e/volumes" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.204345 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.204453 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67076e23-eac4-4e2a-944a-8687060343ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.204475 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.204518 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.204572 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.204590 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4wwx\" (UniqueName: \"kubernetes.io/projected/67076e23-eac4-4e2a-944a-8687060343ad-kube-api-access-c4wwx\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.205824 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67076e23-eac4-4e2a-944a-8687060343ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.209441 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.214886 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.215686 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.222937 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67076e23-eac4-4e2a-944a-8687060343ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.225898 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4wwx\" (UniqueName: \"kubernetes.io/projected/67076e23-eac4-4e2a-944a-8687060343ad-kube-api-access-c4wwx\") pod \"cinder-scheduler-0\" (UID: \"67076e23-eac4-4e2a-944a-8687060343ad\") " pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.298826 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.608365 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6a4a5bce-68e1-448c-a00c-d7fc97e69566","Type":"ContainerStarted","Data":"4a2e92c7de645d020eac4237c63dcb33e4cc1d2fc86cd2ead4143a9b577a15a2"} Dec 04 06:30:59 crc kubenswrapper[4685]: I1204 06:30:59.759847 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 06:30:59 crc kubenswrapper[4685]: W1204 06:30:59.770039 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67076e23_eac4_4e2a_944a_8687060343ad.slice/crio-db51f69b67f79ea5389f5fd99a2451b4ba7243148a30a3cf70c7414cc385bbb3 WatchSource:0}: Error finding container db51f69b67f79ea5389f5fd99a2451b4ba7243148a30a3cf70c7414cc385bbb3: Status 404 returned error can't find the container with id db51f69b67f79ea5389f5fd99a2451b4ba7243148a30a3cf70c7414cc385bbb3 Dec 04 06:31:00 crc kubenswrapper[4685]: I1204 06:31:00.619306 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67076e23-eac4-4e2a-944a-8687060343ad","Type":"ContainerStarted","Data":"eaba4deff82ca2041e01aa42fccf2a882f7f72e81f0ac0279f2a9ffb23ca14d3"} Dec 04 06:31:00 crc kubenswrapper[4685]: I1204 06:31:00.619774 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67076e23-eac4-4e2a-944a-8687060343ad","Type":"ContainerStarted","Data":"db51f69b67f79ea5389f5fd99a2451b4ba7243148a30a3cf70c7414cc385bbb3"} Dec 04 06:31:00 crc kubenswrapper[4685]: I1204 06:31:00.791293 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 06:31:01 crc kubenswrapper[4685]: I1204 06:31:01.632608 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67076e23-eac4-4e2a-944a-8687060343ad","Type":"ContainerStarted","Data":"46f2ce8186f94716e0c26e6b7738b72f93a262c8588c6a5e827d4671208f9bcd"} Dec 04 06:31:01 crc kubenswrapper[4685]: I1204 06:31:01.655630 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.6556085080000003 podStartE2EDuration="3.655608508s" podCreationTimestamp="2025-12-04 06:30:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:01.650131386 +0000 UTC m=+1198.838362161" watchObservedRunningTime="2025-12-04 06:31:01.655608508 +0000 UTC m=+1198.843839293" Dec 04 06:31:02 crc kubenswrapper[4685]: I1204 06:31:02.902783 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6bdfd7d6f6-8slln" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.110301 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-69f76768ff-42jpz"] Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.112995 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.123351 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.123927 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.124837 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.173271 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-69f76768ff-42jpz"] Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.195918 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70d30e39-7d08-4587-90cb-48fc2c868dbc-etc-swift\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196020 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-public-tls-certs\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196058 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d30e39-7d08-4587-90cb-48fc2c868dbc-log-httpd\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196079 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-combined-ca-bundle\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196118 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d30e39-7d08-4587-90cb-48fc2c868dbc-run-httpd\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196180 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-config-data\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196201 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-internal-tls-certs\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.196215 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st789\" (UniqueName: \"kubernetes.io/projected/70d30e39-7d08-4587-90cb-48fc2c868dbc-kube-api-access-st789\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-public-tls-certs\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297661 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d30e39-7d08-4587-90cb-48fc2c868dbc-log-httpd\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297683 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-combined-ca-bundle\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297709 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d30e39-7d08-4587-90cb-48fc2c868dbc-run-httpd\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297765 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-config-data\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297822 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-internal-tls-certs\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297842 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st789\" (UniqueName: \"kubernetes.io/projected/70d30e39-7d08-4587-90cb-48fc2c868dbc-kube-api-access-st789\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.297899 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70d30e39-7d08-4587-90cb-48fc2c868dbc-etc-swift\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.298368 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d30e39-7d08-4587-90cb-48fc2c868dbc-run-httpd\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.299220 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d30e39-7d08-4587-90cb-48fc2c868dbc-log-httpd\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.306174 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70d30e39-7d08-4587-90cb-48fc2c868dbc-etc-swift\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.307542 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-config-data\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.314552 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-combined-ca-bundle\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.315057 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-public-tls-certs\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.319134 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d30e39-7d08-4587-90cb-48fc2c868dbc-internal-tls-certs\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.334300 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st789\" (UniqueName: \"kubernetes.io/projected/70d30e39-7d08-4587-90cb-48fc2c868dbc-kube-api-access-st789\") pod \"swift-proxy-69f76768ff-42jpz\" (UID: \"70d30e39-7d08-4587-90cb-48fc2c868dbc\") " pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.447510 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:03 crc kubenswrapper[4685]: I1204 06:31:03.972969 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-69f76768ff-42jpz"] Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.299535 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.369478 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.370433 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="proxy-httpd" containerID="cri-o://1941db914492dacfb073d1f6d709ad5371ccbb812f95c6e627e286ac180d8f67" gracePeriod=30 Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.370550 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="sg-core" containerID="cri-o://701540091a402585044499802211f8b4eb9b948744e08d903570780edbd2eebc" gracePeriod=30 Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.370604 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-notification-agent" containerID="cri-o://dafb7bf39e28a94dddd392c969e530dae6832d12dc474bb4b856873dcabd113e" gracePeriod=30 Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.369811 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-central-agent" containerID="cri-o://0ed325be65dd9582f5d7cceab1ee66bb9dcf354c57f529af3b0bbae0778e58f9" gracePeriod=30 Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.390697 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.676011 4685 generic.go:334] "Generic (PLEG): container finished" podID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerID="1941db914492dacfb073d1f6d709ad5371ccbb812f95c6e627e286ac180d8f67" exitCode=0 Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.676291 4685 generic.go:334] "Generic (PLEG): container finished" podID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerID="701540091a402585044499802211f8b4eb9b948744e08d903570780edbd2eebc" exitCode=2 Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.676086 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerDied","Data":"1941db914492dacfb073d1f6d709ad5371ccbb812f95c6e627e286ac180d8f67"} Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.676323 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerDied","Data":"701540091a402585044499802211f8b4eb9b948744e08d903570780edbd2eebc"} Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.894719 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-sw96z"] Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.895891 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.906340 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sw96z"] Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.981986 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-r9p2n"] Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.983975 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:04 crc kubenswrapper[4685]: I1204 06:31:04.992783 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r9p2n"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.022508 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0f7b-account-create-update-tjxnm"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.023655 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.025818 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.030722 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhbt7\" (UniqueName: \"kubernetes.io/projected/4c346e92-8b63-4236-81fc-83864447360a-kube-api-access-bhbt7\") pod \"nova-api-db-create-sw96z\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.030755 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c346e92-8b63-4236-81fc-83864447360a-operator-scripts\") pod \"nova-api-db-create-sw96z\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.032354 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0f7b-account-create-update-tjxnm"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.131770 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-operator-scripts\") pod \"nova-api-0f7b-account-create-update-tjxnm\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.131811 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21dbe351-9039-4820-a75d-e06bbaba52f4-operator-scripts\") pod \"nova-cell0-db-create-r9p2n\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.131849 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84rqj\" (UniqueName: \"kubernetes.io/projected/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-kube-api-access-84rqj\") pod \"nova-api-0f7b-account-create-update-tjxnm\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.131974 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhbt7\" (UniqueName: \"kubernetes.io/projected/4c346e92-8b63-4236-81fc-83864447360a-kube-api-access-bhbt7\") pod \"nova-api-db-create-sw96z\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.131998 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c346e92-8b63-4236-81fc-83864447360a-operator-scripts\") pod \"nova-api-db-create-sw96z\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.132022 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w6qt\" (UniqueName: \"kubernetes.io/projected/21dbe351-9039-4820-a75d-e06bbaba52f4-kube-api-access-9w6qt\") pod \"nova-cell0-db-create-r9p2n\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.132986 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c346e92-8b63-4236-81fc-83864447360a-operator-scripts\") pod \"nova-api-db-create-sw96z\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.156954 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhbt7\" (UniqueName: \"kubernetes.io/projected/4c346e92-8b63-4236-81fc-83864447360a-kube-api-access-bhbt7\") pod \"nova-api-db-create-sw96z\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.196297 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-nzzq6"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.197454 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.213956 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3e55-account-create-update-rtbnz"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.215121 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.217933 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.218563 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.223527 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nzzq6"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.229742 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3e55-account-create-update-rtbnz"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.241208 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w6qt\" (UniqueName: \"kubernetes.io/projected/21dbe351-9039-4820-a75d-e06bbaba52f4-kube-api-access-9w6qt\") pod \"nova-cell0-db-create-r9p2n\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.241337 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-operator-scripts\") pod \"nova-api-0f7b-account-create-update-tjxnm\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.241378 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21dbe351-9039-4820-a75d-e06bbaba52f4-operator-scripts\") pod \"nova-cell0-db-create-r9p2n\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.242217 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-operator-scripts\") pod \"nova-api-0f7b-account-create-update-tjxnm\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.241572 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84rqj\" (UniqueName: \"kubernetes.io/projected/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-kube-api-access-84rqj\") pod \"nova-api-0f7b-account-create-update-tjxnm\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.242353 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21dbe351-9039-4820-a75d-e06bbaba52f4-operator-scripts\") pod \"nova-cell0-db-create-r9p2n\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.257790 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84rqj\" (UniqueName: \"kubernetes.io/projected/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-kube-api-access-84rqj\") pod \"nova-api-0f7b-account-create-update-tjxnm\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.272323 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w6qt\" (UniqueName: \"kubernetes.io/projected/21dbe351-9039-4820-a75d-e06bbaba52f4-kube-api-access-9w6qt\") pod \"nova-cell0-db-create-r9p2n\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.311225 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.338222 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.344382 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26e1d9c5-7b1f-41bb-b58b-190488337724-operator-scripts\") pod \"nova-cell0-3e55-account-create-update-rtbnz\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.344445 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mklpr\" (UniqueName: \"kubernetes.io/projected/26e1d9c5-7b1f-41bb-b58b-190488337724-kube-api-access-mklpr\") pod \"nova-cell0-3e55-account-create-update-rtbnz\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.344566 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt44z\" (UniqueName: \"kubernetes.io/projected/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-kube-api-access-tt44z\") pod \"nova-cell1-db-create-nzzq6\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.344619 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-operator-scripts\") pod \"nova-cell1-db-create-nzzq6\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.397320 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-e025-account-create-update-xg92x"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.398676 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.403402 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.424127 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e025-account-create-update-xg92x"] Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.446677 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt44z\" (UniqueName: \"kubernetes.io/projected/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-kube-api-access-tt44z\") pod \"nova-cell1-db-create-nzzq6\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.446758 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-operator-scripts\") pod \"nova-cell1-db-create-nzzq6\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.446796 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26e1d9c5-7b1f-41bb-b58b-190488337724-operator-scripts\") pod \"nova-cell0-3e55-account-create-update-rtbnz\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.446824 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mklpr\" (UniqueName: \"kubernetes.io/projected/26e1d9c5-7b1f-41bb-b58b-190488337724-kube-api-access-mklpr\") pod \"nova-cell0-3e55-account-create-update-rtbnz\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.449358 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-operator-scripts\") pod \"nova-cell1-db-create-nzzq6\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.449853 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26e1d9c5-7b1f-41bb-b58b-190488337724-operator-scripts\") pod \"nova-cell0-3e55-account-create-update-rtbnz\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.468550 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mklpr\" (UniqueName: \"kubernetes.io/projected/26e1d9c5-7b1f-41bb-b58b-190488337724-kube-api-access-mklpr\") pod \"nova-cell0-3e55-account-create-update-rtbnz\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.475824 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt44z\" (UniqueName: \"kubernetes.io/projected/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-kube-api-access-tt44z\") pod \"nova-cell1-db-create-nzzq6\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.550660 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rfff\" (UniqueName: \"kubernetes.io/projected/4d488956-27a8-4cea-8adf-2bc498bf1868-kube-api-access-4rfff\") pod \"nova-cell1-e025-account-create-update-xg92x\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.551040 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d488956-27a8-4cea-8adf-2bc498bf1868-operator-scripts\") pod \"nova-cell1-e025-account-create-update-xg92x\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.558204 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.639700 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.653329 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rfff\" (UniqueName: \"kubernetes.io/projected/4d488956-27a8-4cea-8adf-2bc498bf1868-kube-api-access-4rfff\") pod \"nova-cell1-e025-account-create-update-xg92x\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.653484 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d488956-27a8-4cea-8adf-2bc498bf1868-operator-scripts\") pod \"nova-cell1-e025-account-create-update-xg92x\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.654228 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d488956-27a8-4cea-8adf-2bc498bf1868-operator-scripts\") pod \"nova-cell1-e025-account-create-update-xg92x\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.669327 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rfff\" (UniqueName: \"kubernetes.io/projected/4d488956-27a8-4cea-8adf-2bc498bf1868-kube-api-access-4rfff\") pod \"nova-cell1-e025-account-create-update-xg92x\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.692682 4685 generic.go:334] "Generic (PLEG): container finished" podID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerID="dafb7bf39e28a94dddd392c969e530dae6832d12dc474bb4b856873dcabd113e" exitCode=0 Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.692716 4685 generic.go:334] "Generic (PLEG): container finished" podID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerID="0ed325be65dd9582f5d7cceab1ee66bb9dcf354c57f529af3b0bbae0778e58f9" exitCode=0 Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.692737 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerDied","Data":"dafb7bf39e28a94dddd392c969e530dae6832d12dc474bb4b856873dcabd113e"} Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.692781 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerDied","Data":"0ed325be65dd9582f5d7cceab1ee66bb9dcf354c57f529af3b0bbae0778e58f9"} Dec 04 06:31:05 crc kubenswrapper[4685]: I1204 06:31:05.725590 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:09 crc kubenswrapper[4685]: W1204 06:31:09.035702 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70d30e39_7d08_4587_90cb_48fc2c868dbc.slice/crio-abd7fc22c4663c3ddb71aa04c0160ceebd63f686bd6bb034b4d52cacd9fda12e WatchSource:0}: Error finding container abd7fc22c4663c3ddb71aa04c0160ceebd63f686bd6bb034b4d52cacd9fda12e: Status 404 returned error can't find the container with id abd7fc22c4663c3ddb71aa04c0160ceebd63f686bd6bb034b4d52cacd9fda12e Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.583855 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.671024 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.746238 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6a4a5bce-68e1-448c-a00c-d7fc97e69566","Type":"ContainerStarted","Data":"56ff38ffab36fe428a899ad857b8c958e948bcb5e1ca74addd3783165a714254"} Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.751258 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.751253 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d511b5-5f39-4ea8-84f9-a29f1587368e","Type":"ContainerDied","Data":"b12aa6ba60dd1eb9e639d669ef150feb18186c25dfd352c25f2475064bfc05ec"} Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.751436 4685 scope.go:117] "RemoveContainer" containerID="1941db914492dacfb073d1f6d709ad5371ccbb812f95c6e627e286ac180d8f67" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.754719 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-69f76768ff-42jpz" event={"ID":"70d30e39-7d08-4587-90cb-48fc2c868dbc","Type":"ContainerStarted","Data":"b6cd802447d802581bf7b959ffc96448e74441be129e2674396cb1bdf62f5dc7"} Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.754991 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-69f76768ff-42jpz" event={"ID":"70d30e39-7d08-4587-90cb-48fc2c868dbc","Type":"ContainerStarted","Data":"efd0b7f5b7f3ce5050e4012109a296ddcd4ea63e999dfcd987d492cb8a1e2107"} Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.755036 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-69f76768ff-42jpz" event={"ID":"70d30e39-7d08-4587-90cb-48fc2c868dbc","Type":"ContainerStarted","Data":"abd7fc22c4663c3ddb71aa04c0160ceebd63f686bd6bb034b4d52cacd9fda12e"} Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.755998 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.756028 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.763932 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nzzq6"] Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.769332 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.312361681 podStartE2EDuration="12.769313145s" podCreationTimestamp="2025-12-04 06:30:57 +0000 UTC" firstStartedPulling="2025-12-04 06:30:58.803026895 +0000 UTC m=+1195.991257680" lastFinishedPulling="2025-12-04 06:31:09.259978369 +0000 UTC m=+1206.448209144" observedRunningTime="2025-12-04 06:31:09.763681278 +0000 UTC m=+1206.951912053" watchObservedRunningTime="2025-12-04 06:31:09.769313145 +0000 UTC m=+1206.957543920" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.785965 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-69f76768ff-42jpz" podStartSLOduration=6.7859478079999995 podStartE2EDuration="6.785947808s" podCreationTimestamp="2025-12-04 06:31:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:09.78250131 +0000 UTC m=+1206.970732085" watchObservedRunningTime="2025-12-04 06:31:09.785947808 +0000 UTC m=+1206.974178583" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.788748 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-run-httpd\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.788792 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-config-data\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.788878 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-sg-core-conf-yaml\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.788935 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-scripts\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.789059 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-log-httpd\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.789088 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fgdj\" (UniqueName: \"kubernetes.io/projected/58d511b5-5f39-4ea8-84f9-a29f1587368e-kube-api-access-5fgdj\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.789135 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-combined-ca-bundle\") pod \"58d511b5-5f39-4ea8-84f9-a29f1587368e\" (UID: \"58d511b5-5f39-4ea8-84f9-a29f1587368e\") " Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.790164 4685 scope.go:117] "RemoveContainer" containerID="701540091a402585044499802211f8b4eb9b948744e08d903570780edbd2eebc" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.790727 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.796756 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.803525 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-scripts" (OuterVolumeSpecName: "scripts") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.803927 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.804001 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d511b5-5f39-4ea8-84f9-a29f1587368e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.814206 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d511b5-5f39-4ea8-84f9-a29f1587368e-kube-api-access-5fgdj" (OuterVolumeSpecName: "kube-api-access-5fgdj") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "kube-api-access-5fgdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.859813 4685 scope.go:117] "RemoveContainer" containerID="dafb7bf39e28a94dddd392c969e530dae6832d12dc474bb4b856873dcabd113e" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.872331 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.898613 4685 scope.go:117] "RemoveContainer" containerID="0ed325be65dd9582f5d7cceab1ee66bb9dcf354c57f529af3b0bbae0778e58f9" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.906064 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.906089 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.906098 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fgdj\" (UniqueName: \"kubernetes.io/projected/58d511b5-5f39-4ea8-84f9-a29f1587368e-kube-api-access-5fgdj\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.938942 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:09 crc kubenswrapper[4685]: I1204 06:31:09.964614 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-config-data" (OuterVolumeSpecName: "config-data") pod "58d511b5-5f39-4ea8-84f9-a29f1587368e" (UID: "58d511b5-5f39-4ea8-84f9-a29f1587368e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.008064 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.008096 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d511b5-5f39-4ea8-84f9-a29f1587368e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.143756 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.155947 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.175476 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e025-account-create-update-xg92x"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.208382 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:10 crc kubenswrapper[4685]: E1204 06:31:10.209066 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="proxy-httpd" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209081 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="proxy-httpd" Dec 04 06:31:10 crc kubenswrapper[4685]: E1204 06:31:10.209100 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-central-agent" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209107 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-central-agent" Dec 04 06:31:10 crc kubenswrapper[4685]: E1204 06:31:10.209123 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="sg-core" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209130 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="sg-core" Dec 04 06:31:10 crc kubenswrapper[4685]: E1204 06:31:10.209142 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-notification-agent" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209148 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-notification-agent" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209317 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-notification-agent" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209340 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="sg-core" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209350 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="proxy-httpd" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.209365 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" containerName="ceilometer-central-agent" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.211264 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.218086 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.218295 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.219938 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0f7b-account-create-update-tjxnm"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.235723 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.249121 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r9p2n"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.260467 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sw96z"] Dec 04 06:31:10 crc kubenswrapper[4685]: W1204 06:31:10.262292 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26e1d9c5_7b1f_41bb_b58b_190488337724.slice/crio-dc779e3d73b3e3584638f818f7b79b68ee38df4382d9936cdbc9fe6f1dd30e5d WatchSource:0}: Error finding container dc779e3d73b3e3584638f818f7b79b68ee38df4382d9936cdbc9fe6f1dd30e5d: Status 404 returned error can't find the container with id dc779e3d73b3e3584638f818f7b79b68ee38df4382d9936cdbc9fe6f1dd30e5d Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.269570 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3e55-account-create-update-rtbnz"] Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314479 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-scripts\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314592 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314626 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-config-data\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314642 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crf7q\" (UniqueName: \"kubernetes.io/projected/6890e66b-0287-4564-a42f-073f3ac3ccde-kube-api-access-crf7q\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314671 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-log-httpd\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.314701 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-run-httpd\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.415878 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.416211 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-config-data\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.416320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crf7q\" (UniqueName: \"kubernetes.io/projected/6890e66b-0287-4564-a42f-073f3ac3ccde-kube-api-access-crf7q\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.416450 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-log-httpd\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.416569 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-run-httpd\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.416698 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-scripts\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.416832 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.417146 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-log-httpd\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.417536 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-run-httpd\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.422316 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.423323 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-scripts\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.423652 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.428473 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-config-data\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.433000 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crf7q\" (UniqueName: \"kubernetes.io/projected/6890e66b-0287-4564-a42f-073f3ac3ccde-kube-api-access-crf7q\") pod \"ceilometer-0\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.516231 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.787538 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sw96z" event={"ID":"4c346e92-8b63-4236-81fc-83864447360a","Type":"ContainerStarted","Data":"82af9448da433fdbf3a5b824e316a3c6d90a2ffb7ce5cdbcae25106f0a5fe63a"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.797692 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e025-account-create-update-xg92x" event={"ID":"4d488956-27a8-4cea-8adf-2bc498bf1868","Type":"ContainerStarted","Data":"be101c3dbd1d2db61a2bc85f87fb9d020e653404100f53d25e67d5b7b57231ab"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.809676 4685 generic.go:334] "Generic (PLEG): container finished" podID="baf8cf2f-5cdb-45b0-898a-1c7f57330a86" containerID="62ea205bcc848b1fb70edea493bc2bc1106dacc6b3196e4c3d38f7a24555f1de" exitCode=0 Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.809738 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nzzq6" event={"ID":"baf8cf2f-5cdb-45b0-898a-1c7f57330a86","Type":"ContainerDied","Data":"62ea205bcc848b1fb70edea493bc2bc1106dacc6b3196e4c3d38f7a24555f1de"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.809765 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nzzq6" event={"ID":"baf8cf2f-5cdb-45b0-898a-1c7f57330a86","Type":"ContainerStarted","Data":"ab1f55c303026fb098ba2dd437da8ba2852ae117d4610037e8c92eee36d08bd7"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.831572 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" event={"ID":"26e1d9c5-7b1f-41bb-b58b-190488337724","Type":"ContainerStarted","Data":"5f45cd1834d62289fa089266d9593c98c8f8ddc608e78e7bca16ae023f56bfb9"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.831618 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" event={"ID":"26e1d9c5-7b1f-41bb-b58b-190488337724","Type":"ContainerStarted","Data":"dc779e3d73b3e3584638f818f7b79b68ee38df4382d9936cdbc9fe6f1dd30e5d"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.854648 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r9p2n" event={"ID":"21dbe351-9039-4820-a75d-e06bbaba52f4","Type":"ContainerStarted","Data":"d5be38bcad8d1aaf32f17f50810883c8dae3c6fb4a56e660776127539507d9cc"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.854697 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r9p2n" event={"ID":"21dbe351-9039-4820-a75d-e06bbaba52f4","Type":"ContainerStarted","Data":"75ae84e9c823f53c0347b61fd2c2ad19e0cea460a48d44e2e1f8a196195f77a3"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.855117 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" podStartSLOduration=5.855095118 podStartE2EDuration="5.855095118s" podCreationTimestamp="2025-12-04 06:31:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:10.852699653 +0000 UTC m=+1208.040930428" watchObservedRunningTime="2025-12-04 06:31:10.855095118 +0000 UTC m=+1208.043325893" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.861933 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" event={"ID":"2d0095a3-18ba-4bec-b36a-fd9720d76ddd","Type":"ContainerStarted","Data":"f5fc4688f928b73cc2061273f738cf8a8f70956d624c8ba3e567b84ffd763b17"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.861976 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" event={"ID":"2d0095a3-18ba-4bec-b36a-fd9720d76ddd","Type":"ContainerStarted","Data":"a195f22808990c2d336d0d30298682ab3d098071f0202951520627f187baa68c"} Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.876924 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-r9p2n" podStartSLOduration=6.876903264 podStartE2EDuration="6.876903264s" podCreationTimestamp="2025-12-04 06:31:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:10.8723094 +0000 UTC m=+1208.060540175" watchObservedRunningTime="2025-12-04 06:31:10.876903264 +0000 UTC m=+1208.065134039" Dec 04 06:31:10 crc kubenswrapper[4685]: I1204 06:31:10.909171 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" podStartSLOduration=6.909151348 podStartE2EDuration="6.909151348s" podCreationTimestamp="2025-12-04 06:31:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:10.907956591 +0000 UTC m=+1208.096187366" watchObservedRunningTime="2025-12-04 06:31:10.909151348 +0000 UTC m=+1208.097382123" Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.136303 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d511b5-5f39-4ea8-84f9-a29f1587368e" path="/var/lib/kubelet/pods/58d511b5-5f39-4ea8-84f9-a29f1587368e/volumes" Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.160859 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.871581 4685 generic.go:334] "Generic (PLEG): container finished" podID="21dbe351-9039-4820-a75d-e06bbaba52f4" containerID="d5be38bcad8d1aaf32f17f50810883c8dae3c6fb4a56e660776127539507d9cc" exitCode=0 Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.871623 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r9p2n" event={"ID":"21dbe351-9039-4820-a75d-e06bbaba52f4","Type":"ContainerDied","Data":"d5be38bcad8d1aaf32f17f50810883c8dae3c6fb4a56e660776127539507d9cc"} Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.873478 4685 generic.go:334] "Generic (PLEG): container finished" podID="2d0095a3-18ba-4bec-b36a-fd9720d76ddd" containerID="f5fc4688f928b73cc2061273f738cf8a8f70956d624c8ba3e567b84ffd763b17" exitCode=0 Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.873509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" event={"ID":"2d0095a3-18ba-4bec-b36a-fd9720d76ddd","Type":"ContainerDied","Data":"f5fc4688f928b73cc2061273f738cf8a8f70956d624c8ba3e567b84ffd763b17"} Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.875361 4685 generic.go:334] "Generic (PLEG): container finished" podID="4c346e92-8b63-4236-81fc-83864447360a" containerID="eb76df02727dedb360949a4bd6dcb29986d216ae56c550dc4f2a5c47ab63142d" exitCode=0 Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.875591 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sw96z" event={"ID":"4c346e92-8b63-4236-81fc-83864447360a","Type":"ContainerDied","Data":"eb76df02727dedb360949a4bd6dcb29986d216ae56c550dc4f2a5c47ab63142d"} Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.877236 4685 generic.go:334] "Generic (PLEG): container finished" podID="4d488956-27a8-4cea-8adf-2bc498bf1868" containerID="66a9b48a97f988823569864e648e76d2467cdf26679169fe46b25d93432385e3" exitCode=0 Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.877303 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e025-account-create-update-xg92x" event={"ID":"4d488956-27a8-4cea-8adf-2bc498bf1868","Type":"ContainerDied","Data":"66a9b48a97f988823569864e648e76d2467cdf26679169fe46b25d93432385e3"} Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.878269 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerStarted","Data":"f4c31117af64997b50ce56ebb2ce511041c6f3aef8cc4cae37faca5c5c718c88"} Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.879863 4685 generic.go:334] "Generic (PLEG): container finished" podID="26e1d9c5-7b1f-41bb-b58b-190488337724" containerID="5f45cd1834d62289fa089266d9593c98c8f8ddc608e78e7bca16ae023f56bfb9" exitCode=0 Dec 04 06:31:11 crc kubenswrapper[4685]: I1204 06:31:11.879951 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" event={"ID":"26e1d9c5-7b1f-41bb-b58b-190488337724","Type":"ContainerDied","Data":"5f45cd1834d62289fa089266d9593c98c8f8ddc608e78e7bca16ae023f56bfb9"} Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.368523 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.468132 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt44z\" (UniqueName: \"kubernetes.io/projected/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-kube-api-access-tt44z\") pod \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.468354 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-operator-scripts\") pod \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\" (UID: \"baf8cf2f-5cdb-45b0-898a-1c7f57330a86\") " Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.469446 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "baf8cf2f-5cdb-45b0-898a-1c7f57330a86" (UID: "baf8cf2f-5cdb-45b0-898a-1c7f57330a86"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.475341 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-kube-api-access-tt44z" (OuterVolumeSpecName: "kube-api-access-tt44z") pod "baf8cf2f-5cdb-45b0-898a-1c7f57330a86" (UID: "baf8cf2f-5cdb-45b0-898a-1c7f57330a86"). InnerVolumeSpecName "kube-api-access-tt44z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.570945 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt44z\" (UniqueName: \"kubernetes.io/projected/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-kube-api-access-tt44z\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.570980 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf8cf2f-5cdb-45b0-898a-1c7f57330a86-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.788306 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.889384 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nzzq6" event={"ID":"baf8cf2f-5cdb-45b0-898a-1c7f57330a86","Type":"ContainerDied","Data":"ab1f55c303026fb098ba2dd437da8ba2852ae117d4610037e8c92eee36d08bd7"} Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.889501 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab1f55c303026fb098ba2dd437da8ba2852ae117d4610037e8c92eee36d08bd7" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.889515 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nzzq6" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.891543 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerStarted","Data":"4c955b58df2272105beea77ff75490bc2c89199b9aa522e44ed715801e3a0336"} Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.891587 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerStarted","Data":"00f7e07c9408537babec74f8e71ae297de9dbc26d4db6b01144b9b9c36735f0b"} Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.904626 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6bdfd7d6f6-8slln" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 06:31:12 crc kubenswrapper[4685]: I1204 06:31:12.904762 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.330447 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.390100 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhbt7\" (UniqueName: \"kubernetes.io/projected/4c346e92-8b63-4236-81fc-83864447360a-kube-api-access-bhbt7\") pod \"4c346e92-8b63-4236-81fc-83864447360a\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.390350 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c346e92-8b63-4236-81fc-83864447360a-operator-scripts\") pod \"4c346e92-8b63-4236-81fc-83864447360a\" (UID: \"4c346e92-8b63-4236-81fc-83864447360a\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.391984 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c346e92-8b63-4236-81fc-83864447360a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c346e92-8b63-4236-81fc-83864447360a" (UID: "4c346e92-8b63-4236-81fc-83864447360a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.469431 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c346e92-8b63-4236-81fc-83864447360a-kube-api-access-bhbt7" (OuterVolumeSpecName: "kube-api-access-bhbt7") pod "4c346e92-8b63-4236-81fc-83864447360a" (UID: "4c346e92-8b63-4236-81fc-83864447360a"). InnerVolumeSpecName "kube-api-access-bhbt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.494249 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c346e92-8b63-4236-81fc-83864447360a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.494282 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhbt7\" (UniqueName: \"kubernetes.io/projected/4c346e92-8b63-4236-81fc-83864447360a-kube-api-access-bhbt7\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.522690 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.524957 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.575271 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.597882 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84rqj\" (UniqueName: \"kubernetes.io/projected/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-kube-api-access-84rqj\") pod \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.598012 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w6qt\" (UniqueName: \"kubernetes.io/projected/21dbe351-9039-4820-a75d-e06bbaba52f4-kube-api-access-9w6qt\") pod \"21dbe351-9039-4820-a75d-e06bbaba52f4\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.598041 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21dbe351-9039-4820-a75d-e06bbaba52f4-operator-scripts\") pod \"21dbe351-9039-4820-a75d-e06bbaba52f4\" (UID: \"21dbe351-9039-4820-a75d-e06bbaba52f4\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.598100 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-operator-scripts\") pod \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\" (UID: \"2d0095a3-18ba-4bec-b36a-fd9720d76ddd\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.598909 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d0095a3-18ba-4bec-b36a-fd9720d76ddd" (UID: "2d0095a3-18ba-4bec-b36a-fd9720d76ddd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.615757 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21dbe351-9039-4820-a75d-e06bbaba52f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "21dbe351-9039-4820-a75d-e06bbaba52f4" (UID: "21dbe351-9039-4820-a75d-e06bbaba52f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.625228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21dbe351-9039-4820-a75d-e06bbaba52f4-kube-api-access-9w6qt" (OuterVolumeSpecName: "kube-api-access-9w6qt") pod "21dbe351-9039-4820-a75d-e06bbaba52f4" (UID: "21dbe351-9039-4820-a75d-e06bbaba52f4"). InnerVolumeSpecName "kube-api-access-9w6qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.627803 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-kube-api-access-84rqj" (OuterVolumeSpecName: "kube-api-access-84rqj") pod "2d0095a3-18ba-4bec-b36a-fd9720d76ddd" (UID: "2d0095a3-18ba-4bec-b36a-fd9720d76ddd"). InnerVolumeSpecName "kube-api-access-84rqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.641934 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.702488 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26e1d9c5-7b1f-41bb-b58b-190488337724-operator-scripts\") pod \"26e1d9c5-7b1f-41bb-b58b-190488337724\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.702605 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mklpr\" (UniqueName: \"kubernetes.io/projected/26e1d9c5-7b1f-41bb-b58b-190488337724-kube-api-access-mklpr\") pod \"26e1d9c5-7b1f-41bb-b58b-190488337724\" (UID: \"26e1d9c5-7b1f-41bb-b58b-190488337724\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.702725 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rfff\" (UniqueName: \"kubernetes.io/projected/4d488956-27a8-4cea-8adf-2bc498bf1868-kube-api-access-4rfff\") pod \"4d488956-27a8-4cea-8adf-2bc498bf1868\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.702768 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d488956-27a8-4cea-8adf-2bc498bf1868-operator-scripts\") pod \"4d488956-27a8-4cea-8adf-2bc498bf1868\" (UID: \"4d488956-27a8-4cea-8adf-2bc498bf1868\") " Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.703107 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w6qt\" (UniqueName: \"kubernetes.io/projected/21dbe351-9039-4820-a75d-e06bbaba52f4-kube-api-access-9w6qt\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.703118 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21dbe351-9039-4820-a75d-e06bbaba52f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.703129 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.703140 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84rqj\" (UniqueName: \"kubernetes.io/projected/2d0095a3-18ba-4bec-b36a-fd9720d76ddd-kube-api-access-84rqj\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.703630 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d488956-27a8-4cea-8adf-2bc498bf1868-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d488956-27a8-4cea-8adf-2bc498bf1868" (UID: "4d488956-27a8-4cea-8adf-2bc498bf1868"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.703737 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e1d9c5-7b1f-41bb-b58b-190488337724-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "26e1d9c5-7b1f-41bb-b58b-190488337724" (UID: "26e1d9c5-7b1f-41bb-b58b-190488337724"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.708126 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e1d9c5-7b1f-41bb-b58b-190488337724-kube-api-access-mklpr" (OuterVolumeSpecName: "kube-api-access-mklpr") pod "26e1d9c5-7b1f-41bb-b58b-190488337724" (UID: "26e1d9c5-7b1f-41bb-b58b-190488337724"). InnerVolumeSpecName "kube-api-access-mklpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.708709 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d488956-27a8-4cea-8adf-2bc498bf1868-kube-api-access-4rfff" (OuterVolumeSpecName: "kube-api-access-4rfff") pod "4d488956-27a8-4cea-8adf-2bc498bf1868" (UID: "4d488956-27a8-4cea-8adf-2bc498bf1868"). InnerVolumeSpecName "kube-api-access-4rfff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.806132 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rfff\" (UniqueName: \"kubernetes.io/projected/4d488956-27a8-4cea-8adf-2bc498bf1868-kube-api-access-4rfff\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.806179 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d488956-27a8-4cea-8adf-2bc498bf1868-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.806193 4685 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26e1d9c5-7b1f-41bb-b58b-190488337724-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.806204 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mklpr\" (UniqueName: \"kubernetes.io/projected/26e1d9c5-7b1f-41bb-b58b-190488337724-kube-api-access-mklpr\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.914002 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" event={"ID":"26e1d9c5-7b1f-41bb-b58b-190488337724","Type":"ContainerDied","Data":"dc779e3d73b3e3584638f818f7b79b68ee38df4382d9936cdbc9fe6f1dd30e5d"} Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.914051 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc779e3d73b3e3584638f818f7b79b68ee38df4382d9936cdbc9fe6f1dd30e5d" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.914057 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e55-account-create-update-rtbnz" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.915848 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r9p2n" event={"ID":"21dbe351-9039-4820-a75d-e06bbaba52f4","Type":"ContainerDied","Data":"75ae84e9c823f53c0347b61fd2c2ad19e0cea460a48d44e2e1f8a196195f77a3"} Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.915879 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75ae84e9c823f53c0347b61fd2c2ad19e0cea460a48d44e2e1f8a196195f77a3" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.915961 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r9p2n" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.921709 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" event={"ID":"2d0095a3-18ba-4bec-b36a-fd9720d76ddd","Type":"ContainerDied","Data":"a195f22808990c2d336d0d30298682ab3d098071f0202951520627f187baa68c"} Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.921754 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a195f22808990c2d336d0d30298682ab3d098071f0202951520627f187baa68c" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.921727 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f7b-account-create-update-tjxnm" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.929153 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sw96z" event={"ID":"4c346e92-8b63-4236-81fc-83864447360a","Type":"ContainerDied","Data":"82af9448da433fdbf3a5b824e316a3c6d90a2ffb7ce5cdbcae25106f0a5fe63a"} Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.929185 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sw96z" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.929198 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82af9448da433fdbf3a5b824e316a3c6d90a2ffb7ce5cdbcae25106f0a5fe63a" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.931028 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e025-account-create-update-xg92x" event={"ID":"4d488956-27a8-4cea-8adf-2bc498bf1868","Type":"ContainerDied","Data":"be101c3dbd1d2db61a2bc85f87fb9d020e653404100f53d25e67d5b7b57231ab"} Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.931059 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be101c3dbd1d2db61a2bc85f87fb9d020e653404100f53d25e67d5b7b57231ab" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.931138 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e025-account-create-update-xg92x" Dec 04 06:31:13 crc kubenswrapper[4685]: I1204 06:31:13.933123 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerStarted","Data":"10373be4a7524333bfeca9e68393da70d3592ea74d0c0234ee92d116029c0d8a"} Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.666210 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-chrlf"] Dec 04 06:31:15 crc kubenswrapper[4685]: E1204 06:31:15.667039 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0095a3-18ba-4bec-b36a-fd9720d76ddd" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667052 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0095a3-18ba-4bec-b36a-fd9720d76ddd" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: E1204 06:31:15.667075 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21dbe351-9039-4820-a75d-e06bbaba52f4" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667082 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="21dbe351-9039-4820-a75d-e06bbaba52f4" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: E1204 06:31:15.667093 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e1d9c5-7b1f-41bb-b58b-190488337724" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667102 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e1d9c5-7b1f-41bb-b58b-190488337724" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: E1204 06:31:15.667114 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d488956-27a8-4cea-8adf-2bc498bf1868" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667120 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d488956-27a8-4cea-8adf-2bc498bf1868" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: E1204 06:31:15.667134 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf8cf2f-5cdb-45b0-898a-1c7f57330a86" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667140 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf8cf2f-5cdb-45b0-898a-1c7f57330a86" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: E1204 06:31:15.667160 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c346e92-8b63-4236-81fc-83864447360a" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667166 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c346e92-8b63-4236-81fc-83864447360a" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667313 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e1d9c5-7b1f-41bb-b58b-190488337724" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667325 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="21dbe351-9039-4820-a75d-e06bbaba52f4" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667340 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d0095a3-18ba-4bec-b36a-fd9720d76ddd" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667352 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d488956-27a8-4cea-8adf-2bc498bf1868" containerName="mariadb-account-create-update" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667358 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c346e92-8b63-4236-81fc-83864447360a" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.667374 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf8cf2f-5cdb-45b0-898a-1c7f57330a86" containerName="mariadb-database-create" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.673068 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.678862 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.679184 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tpmkh" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.679345 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.680120 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-chrlf"] Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.739751 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdzfd\" (UniqueName: \"kubernetes.io/projected/18ec8793-df06-4f80-9ddc-148283651ef9-kube-api-access-gdzfd\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.739804 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-config-data\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.739835 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-scripts\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.739950 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.841048 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.841137 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdzfd\" (UniqueName: \"kubernetes.io/projected/18ec8793-df06-4f80-9ddc-148283651ef9-kube-api-access-gdzfd\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.841169 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-config-data\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.841193 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-scripts\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.850179 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-scripts\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.850304 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.850393 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-config-data\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.859838 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdzfd\" (UniqueName: \"kubernetes.io/projected/18ec8793-df06-4f80-9ddc-148283651ef9-kube-api-access-gdzfd\") pod \"nova-cell0-conductor-db-sync-chrlf\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.978609 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerStarted","Data":"0fa70c6b42fb93d30312e30905442b468d19132414bd9ead9ae1e5492ec769ee"} Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.978765 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-central-agent" containerID="cri-o://00f7e07c9408537babec74f8e71ae297de9dbc26d4db6b01144b9b9c36735f0b" gracePeriod=30 Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.979001 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.979250 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="proxy-httpd" containerID="cri-o://0fa70c6b42fb93d30312e30905442b468d19132414bd9ead9ae1e5492ec769ee" gracePeriod=30 Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.979292 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="sg-core" containerID="cri-o://10373be4a7524333bfeca9e68393da70d3592ea74d0c0234ee92d116029c0d8a" gracePeriod=30 Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.979325 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-notification-agent" containerID="cri-o://4c955b58df2272105beea77ff75490bc2c89199b9aa522e44ed715801e3a0336" gracePeriod=30 Dec 04 06:31:15 crc kubenswrapper[4685]: I1204 06:31:15.991799 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.008550 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.252159975 podStartE2EDuration="6.008531884s" podCreationTimestamp="2025-12-04 06:31:10 +0000 UTC" firstStartedPulling="2025-12-04 06:31:11.189784438 +0000 UTC m=+1208.378015203" lastFinishedPulling="2025-12-04 06:31:14.946156337 +0000 UTC m=+1212.134387112" observedRunningTime="2025-12-04 06:31:16.008508253 +0000 UTC m=+1213.196739028" watchObservedRunningTime="2025-12-04 06:31:16.008531884 +0000 UTC m=+1213.196762669" Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.575487 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-chrlf"] Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.989185 4685 generic.go:334] "Generic (PLEG): container finished" podID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerID="0fa70c6b42fb93d30312e30905442b468d19132414bd9ead9ae1e5492ec769ee" exitCode=0 Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.989506 4685 generic.go:334] "Generic (PLEG): container finished" podID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerID="10373be4a7524333bfeca9e68393da70d3592ea74d0c0234ee92d116029c0d8a" exitCode=2 Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.989516 4685 generic.go:334] "Generic (PLEG): container finished" podID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerID="4c955b58df2272105beea77ff75490bc2c89199b9aa522e44ed715801e3a0336" exitCode=0 Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.989274 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerDied","Data":"0fa70c6b42fb93d30312e30905442b468d19132414bd9ead9ae1e5492ec769ee"} Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.989588 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerDied","Data":"10373be4a7524333bfeca9e68393da70d3592ea74d0c0234ee92d116029c0d8a"} Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.989603 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerDied","Data":"4c955b58df2272105beea77ff75490bc2c89199b9aa522e44ed715801e3a0336"} Dec 04 06:31:16 crc kubenswrapper[4685]: I1204 06:31:16.990868 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-chrlf" event={"ID":"18ec8793-df06-4f80-9ddc-148283651ef9","Type":"ContainerStarted","Data":"9525921a559fadfac21db3f7b5c56f464f5dfa73b5860a75ca837ead3eaed8c2"} Dec 04 06:31:18 crc kubenswrapper[4685]: I1204 06:31:18.454598 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:18 crc kubenswrapper[4685]: I1204 06:31:18.462379 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-69f76768ff-42jpz" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.068275 4685 generic.go:334] "Generic (PLEG): container finished" podID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerID="252f6828706c15e58fcb7a3b491ecee0a37c85fd571fbefe50fcd5492a05e612" exitCode=137 Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.069019 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bdfd7d6f6-8slln" event={"ID":"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8","Type":"ContainerDied","Data":"252f6828706c15e58fcb7a3b491ecee0a37c85fd571fbefe50fcd5492a05e612"} Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.069059 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bdfd7d6f6-8slln" event={"ID":"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8","Type":"ContainerDied","Data":"9e9c684185854ea65eeabdfa138d41f84665bb035420a994e2d88229d118c302"} Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.069074 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e9c684185854ea65eeabdfa138d41f84665bb035420a994e2d88229d118c302" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.073770 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135517 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-combined-ca-bundle\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135592 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8cjl\" (UniqueName: \"kubernetes.io/projected/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-kube-api-access-h8cjl\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135635 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-config-data\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135682 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-tls-certs\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135718 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-secret-key\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135752 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-logs\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.135816 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-scripts\") pod \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\" (UID: \"6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8\") " Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.137322 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-logs" (OuterVolumeSpecName: "logs") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.160213 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-kube-api-access-h8cjl" (OuterVolumeSpecName: "kube-api-access-h8cjl") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "kube-api-access-h8cjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.164934 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-config-data" (OuterVolumeSpecName: "config-data") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.172015 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.197371 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.212140 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.232208 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-scripts" (OuterVolumeSpecName: "scripts") pod "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" (UID: "6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.237916 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.238119 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.238218 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8cjl\" (UniqueName: \"kubernetes.io/projected/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-kube-api-access-h8cjl\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.238314 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.238391 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.238497 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:20 crc kubenswrapper[4685]: I1204 06:31:20.238575 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:21 crc kubenswrapper[4685]: I1204 06:31:21.076125 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bdfd7d6f6-8slln" Dec 04 06:31:21 crc kubenswrapper[4685]: I1204 06:31:21.139473 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bdfd7d6f6-8slln"] Dec 04 06:31:21 crc kubenswrapper[4685]: I1204 06:31:21.157322 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6bdfd7d6f6-8slln"] Dec 04 06:31:22 crc kubenswrapper[4685]: I1204 06:31:22.104446 4685 generic.go:334] "Generic (PLEG): container finished" podID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerID="00f7e07c9408537babec74f8e71ae297de9dbc26d4db6b01144b9b9c36735f0b" exitCode=0 Dec 04 06:31:22 crc kubenswrapper[4685]: I1204 06:31:22.104490 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerDied","Data":"00f7e07c9408537babec74f8e71ae297de9dbc26d4db6b01144b9b9c36735f0b"} Dec 04 06:31:23 crc kubenswrapper[4685]: I1204 06:31:23.138354 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" path="/var/lib/kubelet/pods/6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8/volumes" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.559593 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.679621 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-config-data\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.680149 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-combined-ca-bundle\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.680191 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crf7q\" (UniqueName: \"kubernetes.io/projected/6890e66b-0287-4564-a42f-073f3ac3ccde-kube-api-access-crf7q\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.680264 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-scripts\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.680356 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-log-httpd\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.680384 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-sg-core-conf-yaml\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.680498 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-run-httpd\") pod \"6890e66b-0287-4564-a42f-073f3ac3ccde\" (UID: \"6890e66b-0287-4564-a42f-073f3ac3ccde\") " Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.681301 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.681612 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.689932 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6890e66b-0287-4564-a42f-073f3ac3ccde-kube-api-access-crf7q" (OuterVolumeSpecName: "kube-api-access-crf7q") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "kube-api-access-crf7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.703161 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.703529 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-log" containerID="cri-o://7a08f48c3a2e5031799c652836a2e4db6dbf7eb4a4f1b49abcb9e0575c771574" gracePeriod=30 Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.703658 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-scripts" (OuterVolumeSpecName: "scripts") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.703695 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-httpd" containerID="cri-o://8a776e43c26649414b0dc2963a340b49b052b633b18c81ea218571f27a43597a" gracePeriod=30 Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.743947 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.783076 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crf7q\" (UniqueName: \"kubernetes.io/projected/6890e66b-0287-4564-a42f-073f3ac3ccde-kube-api-access-crf7q\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.783109 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.783119 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.783129 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.783136 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6890e66b-0287-4564-a42f-073f3ac3ccde-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.789771 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.807851 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-config-data" (OuterVolumeSpecName: "config-data") pod "6890e66b-0287-4564-a42f-073f3ac3ccde" (UID: "6890e66b-0287-4564-a42f-073f3ac3ccde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.885235 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:25 crc kubenswrapper[4685]: I1204 06:31:25.885265 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6890e66b-0287-4564-a42f-073f3ac3ccde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.164383 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-chrlf" event={"ID":"18ec8793-df06-4f80-9ddc-148283651ef9","Type":"ContainerStarted","Data":"d3789629a7d33764b1db5b73e19ab06c1ea82860b04a1c59b836bba229fcfaf9"} Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.166289 4685 generic.go:334] "Generic (PLEG): container finished" podID="5d114929-218b-465a-9116-ed8144ed0636" containerID="7a08f48c3a2e5031799c652836a2e4db6dbf7eb4a4f1b49abcb9e0575c771574" exitCode=143 Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.166373 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5d114929-218b-465a-9116-ed8144ed0636","Type":"ContainerDied","Data":"7a08f48c3a2e5031799c652836a2e4db6dbf7eb4a4f1b49abcb9e0575c771574"} Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.168814 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6890e66b-0287-4564-a42f-073f3ac3ccde","Type":"ContainerDied","Data":"f4c31117af64997b50ce56ebb2ce511041c6f3aef8cc4cae37faca5c5c718c88"} Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.168853 4685 scope.go:117] "RemoveContainer" containerID="0fa70c6b42fb93d30312e30905442b468d19132414bd9ead9ae1e5492ec769ee" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.168924 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.201593 4685 scope.go:117] "RemoveContainer" containerID="10373be4a7524333bfeca9e68393da70d3592ea74d0c0234ee92d116029c0d8a" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.226367 4685 scope.go:117] "RemoveContainer" containerID="4c955b58df2272105beea77ff75490bc2c89199b9aa522e44ed715801e3a0336" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.237436 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-chrlf" podStartSLOduration=2.609663281 podStartE2EDuration="11.237390601s" podCreationTimestamp="2025-12-04 06:31:15 +0000 UTC" firstStartedPulling="2025-12-04 06:31:16.568241104 +0000 UTC m=+1213.756471879" lastFinishedPulling="2025-12-04 06:31:25.195968424 +0000 UTC m=+1222.384199199" observedRunningTime="2025-12-04 06:31:26.213138418 +0000 UTC m=+1223.401369193" watchObservedRunningTime="2025-12-04 06:31:26.237390601 +0000 UTC m=+1223.425621386" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.258116 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.267799 4685 scope.go:117] "RemoveContainer" containerID="00f7e07c9408537babec74f8e71ae297de9dbc26d4db6b01144b9b9c36735f0b" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.285766 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.289647 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:26 crc kubenswrapper[4685]: E1204 06:31:26.290448 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon-log" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.290558 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon-log" Dec 04 06:31:26 crc kubenswrapper[4685]: E1204 06:31:26.290657 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-central-agent" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.290733 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-central-agent" Dec 04 06:31:26 crc kubenswrapper[4685]: E1204 06:31:26.290804 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-notification-agent" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.290854 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-notification-agent" Dec 04 06:31:26 crc kubenswrapper[4685]: E1204 06:31:26.290908 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="sg-core" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.290961 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="sg-core" Dec 04 06:31:26 crc kubenswrapper[4685]: E1204 06:31:26.291022 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291078 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" Dec 04 06:31:26 crc kubenswrapper[4685]: E1204 06:31:26.291135 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="proxy-httpd" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291187 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="proxy-httpd" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291457 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291555 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ccb3333-9d7f-45e9-8209-ff5e2ab8eeb8" containerName="horizon-log" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291635 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="sg-core" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291706 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-central-agent" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291787 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="ceilometer-notification-agent" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.291859 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" containerName="proxy-httpd" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.295382 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.297159 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.299979 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.300236 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.396532 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-config-data\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.396880 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-log-httpd\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.397017 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-scripts\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.397173 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-run-httpd\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.397313 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxkf4\" (UniqueName: \"kubernetes.io/projected/f41de5f3-3313-4d6b-80d8-3337a5aca42c-kube-api-access-bxkf4\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.397381 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.397420 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.501224 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-log-httpd\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.501282 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-scripts\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.501318 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-run-httpd\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.501373 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxkf4\" (UniqueName: \"kubernetes.io/projected/f41de5f3-3313-4d6b-80d8-3337a5aca42c-kube-api-access-bxkf4\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.501431 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.501458 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.502145 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-run-httpd\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.502254 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-config-data\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.502966 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-log-httpd\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.506717 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.507565 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-scripts\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.519653 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-config-data\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.521583 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.524834 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxkf4\" (UniqueName: \"kubernetes.io/projected/f41de5f3-3313-4d6b-80d8-3337a5aca42c-kube-api-access-bxkf4\") pod \"ceilometer-0\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.616133 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.621266 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.621694 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-log" containerID="cri-o://194e09b1df424174a9f3f618498a0018145f01d76d1843da7d57f62bc9c0f61d" gracePeriod=30 Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.621733 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-httpd" containerID="cri-o://af51672273a18d10e56be9de171cfdd540c1fb4d658fcd9f511decfd0008f671" gracePeriod=30 Dec 04 06:31:26 crc kubenswrapper[4685]: I1204 06:31:26.874586 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:27 crc kubenswrapper[4685]: I1204 06:31:27.141573 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6890e66b-0287-4564-a42f-073f3ac3ccde" path="/var/lib/kubelet/pods/6890e66b-0287-4564-a42f-073f3ac3ccde/volumes" Dec 04 06:31:27 crc kubenswrapper[4685]: I1204 06:31:27.183884 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerStarted","Data":"99530c131eeb7f9e41609124f2463291c77584dfc6054188c17db31e1eab1f2b"} Dec 04 06:31:27 crc kubenswrapper[4685]: I1204 06:31:27.190943 4685 generic.go:334] "Generic (PLEG): container finished" podID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerID="194e09b1df424174a9f3f618498a0018145f01d76d1843da7d57f62bc9c0f61d" exitCode=143 Dec 04 06:31:27 crc kubenswrapper[4685]: I1204 06:31:27.191867 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34","Type":"ContainerDied","Data":"194e09b1df424174a9f3f618498a0018145f01d76d1843da7d57f62bc9c0f61d"} Dec 04 06:31:28 crc kubenswrapper[4685]: I1204 06:31:28.095827 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:28 crc kubenswrapper[4685]: I1204 06:31:28.201595 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerStarted","Data":"60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812"} Dec 04 06:31:28 crc kubenswrapper[4685]: I1204 06:31:28.201648 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerStarted","Data":"efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88"} Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.214857 4685 generic.go:334] "Generic (PLEG): container finished" podID="5d114929-218b-465a-9116-ed8144ed0636" containerID="8a776e43c26649414b0dc2963a340b49b052b633b18c81ea218571f27a43597a" exitCode=0 Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.214930 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5d114929-218b-465a-9116-ed8144ed0636","Type":"ContainerDied","Data":"8a776e43c26649414b0dc2963a340b49b052b633b18c81ea218571f27a43597a"} Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.219394 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerStarted","Data":"db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912"} Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.948577 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.985807 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-public-tls-certs\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.986250 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-combined-ca-bundle\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.986302 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgpfn\" (UniqueName: \"kubernetes.io/projected/5d114929-218b-465a-9116-ed8144ed0636-kube-api-access-bgpfn\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.986347 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-logs\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.988350 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-config-data\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.988602 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-scripts\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.988674 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.988730 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-httpd-run\") pod \"5d114929-218b-465a-9116-ed8144ed0636\" (UID: \"5d114929-218b-465a-9116-ed8144ed0636\") " Dec 04 06:31:29 crc kubenswrapper[4685]: I1204 06:31:29.990710 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-logs" (OuterVolumeSpecName: "logs") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.001516 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.001888 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-scripts" (OuterVolumeSpecName: "scripts") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.002562 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d114929-218b-465a-9116-ed8144ed0636-kube-api-access-bgpfn" (OuterVolumeSpecName: "kube-api-access-bgpfn") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "kube-api-access-bgpfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.006120 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.017830 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.106927 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.106984 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.106995 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d114929-218b-465a-9116-ed8144ed0636-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.107004 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgpfn\" (UniqueName: \"kubernetes.io/projected/5d114929-218b-465a-9116-ed8144ed0636-kube-api-access-bgpfn\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.108738 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.164643 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-config-data" (OuterVolumeSpecName: "config-data") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.192599 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5d114929-218b-465a-9116-ed8144ed0636" (UID: "5d114929-218b-465a-9116-ed8144ed0636"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.213067 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.213099 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.213109 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d114929-218b-465a-9116-ed8144ed0636-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.219121 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.314720 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.330051 4685 generic.go:334] "Generic (PLEG): container finished" podID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerID="af51672273a18d10e56be9de171cfdd540c1fb4d658fcd9f511decfd0008f671" exitCode=0 Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.330136 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34","Type":"ContainerDied","Data":"af51672273a18d10e56be9de171cfdd540c1fb4d658fcd9f511decfd0008f671"} Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.351684 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5d114929-218b-465a-9116-ed8144ed0636","Type":"ContainerDied","Data":"131407d07c03df3460693da013ab6d63e82d2b24d4c73956e55aacee1f753362"} Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.351740 4685 scope.go:117] "RemoveContainer" containerID="8a776e43c26649414b0dc2963a340b49b052b633b18c81ea218571f27a43597a" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.351913 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.426244 4685 scope.go:117] "RemoveContainer" containerID="7a08f48c3a2e5031799c652836a2e4db6dbf7eb4a4f1b49abcb9e0575c771574" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.435135 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.459823 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.474018 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.494561 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:31:30 crc kubenswrapper[4685]: E1204 06:31:30.494963 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-log" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.494980 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-log" Dec 04 06:31:30 crc kubenswrapper[4685]: E1204 06:31:30.494994 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-httpd" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495000 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-httpd" Dec 04 06:31:30 crc kubenswrapper[4685]: E1204 06:31:30.495015 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-log" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495021 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-log" Dec 04 06:31:30 crc kubenswrapper[4685]: E1204 06:31:30.495051 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-httpd" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495056 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-httpd" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495215 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-httpd" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495230 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-log" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495244 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" containerName="glance-log" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.495256 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d114929-218b-465a-9116-ed8144ed0636" containerName="glance-httpd" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.496176 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.500529 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.500881 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.509606 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518107 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-config-data\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518172 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-internal-tls-certs\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518248 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-combined-ca-bundle\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518284 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-logs\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518322 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518345 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-httpd-run\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518445 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-scripts\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.518481 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7z7d\" (UniqueName: \"kubernetes.io/projected/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-kube-api-access-c7z7d\") pod \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\" (UID: \"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34\") " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.519773 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-logs" (OuterVolumeSpecName: "logs") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.520194 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.526438 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-kube-api-access-c7z7d" (OuterVolumeSpecName: "kube-api-access-c7z7d") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "kube-api-access-c7z7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.532569 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.534242 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-scripts" (OuterVolumeSpecName: "scripts") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.586974 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.591565 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.605314 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-config-data" (OuterVolumeSpecName: "config-data") pod "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" (UID: "8d1b98dd-9b1f-474d-a6b0-7d0385db6a34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620303 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c17edd9-e27f-4397-bfad-baeb4684038e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620359 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c17edd9-e27f-4397-bfad-baeb4684038e-logs\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620389 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620463 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620488 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620505 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620521 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620549 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vsng\" (UniqueName: \"kubernetes.io/projected/8c17edd9-e27f-4397-bfad-baeb4684038e-kube-api-access-4vsng\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620623 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620636 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620645 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620653 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7z7d\" (UniqueName: \"kubernetes.io/projected/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-kube-api-access-c7z7d\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620661 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620669 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620677 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.620684 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.648207 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722602 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c17edd9-e27f-4397-bfad-baeb4684038e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722655 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c17edd9-e27f-4397-bfad-baeb4684038e-logs\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722688 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722760 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722782 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722798 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722816 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722846 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vsng\" (UniqueName: \"kubernetes.io/projected/8c17edd9-e27f-4397-bfad-baeb4684038e-kube-api-access-4vsng\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.722894 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.723235 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c17edd9-e27f-4397-bfad-baeb4684038e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.723680 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c17edd9-e27f-4397-bfad-baeb4684038e-logs\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.724106 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.728764 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.729462 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.730003 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.735633 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c17edd9-e27f-4397-bfad-baeb4684038e-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.747169 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vsng\" (UniqueName: \"kubernetes.io/projected/8c17edd9-e27f-4397-bfad-baeb4684038e-kube-api-access-4vsng\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.794523 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8c17edd9-e27f-4397-bfad-baeb4684038e\") " pod="openstack/glance-default-external-api-0" Dec 04 06:31:30 crc kubenswrapper[4685]: I1204 06:31:30.831286 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.134386 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d114929-218b-465a-9116-ed8144ed0636" path="/var/lib/kubelet/pods/5d114929-218b-465a-9116-ed8144ed0636/volumes" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.361307 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.361500 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d1b98dd-9b1f-474d-a6b0-7d0385db6a34","Type":"ContainerDied","Data":"bda258831aeb73f79d09ceafebd3635a2181f184109f19edc1c368eb9c2dcda7"} Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.361565 4685 scope.go:117] "RemoveContainer" containerID="af51672273a18d10e56be9de171cfdd540c1fb4d658fcd9f511decfd0008f671" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.367488 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerStarted","Data":"86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f"} Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.367637 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-central-agent" containerID="cri-o://efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88" gracePeriod=30 Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.367712 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="proxy-httpd" containerID="cri-o://86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f" gracePeriod=30 Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.367729 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-notification-agent" containerID="cri-o://60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812" gracePeriod=30 Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.367734 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.367706 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="sg-core" containerID="cri-o://db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912" gracePeriod=30 Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.390938 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.402604 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.415719 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.415783 4685 scope.go:117] "RemoveContainer" containerID="194e09b1df424174a9f3f618498a0018145f01d76d1843da7d57f62bc9c0f61d" Dec 04 06:31:31 crc kubenswrapper[4685]: W1204 06:31:31.426460 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c17edd9_e27f_4397_bfad_baeb4684038e.slice/crio-4c4b7b44fa5c36b3021c50aee7069d6ab4eb955a8ce5fc836d7c16481dd7c291 WatchSource:0}: Error finding container 4c4b7b44fa5c36b3021c50aee7069d6ab4eb955a8ce5fc836d7c16481dd7c291: Status 404 returned error can't find the container with id 4c4b7b44fa5c36b3021c50aee7069d6ab4eb955a8ce5fc836d7c16481dd7c291 Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.429121 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.430818 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.438851 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.438834 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.362892997 podStartE2EDuration="5.438811547s" podCreationTimestamp="2025-12-04 06:31:26 +0000 UTC" firstStartedPulling="2025-12-04 06:31:26.885682339 +0000 UTC m=+1224.073913114" lastFinishedPulling="2025-12-04 06:31:29.961600889 +0000 UTC m=+1227.149831664" observedRunningTime="2025-12-04 06:31:31.42207275 +0000 UTC m=+1228.610303525" watchObservedRunningTime="2025-12-04 06:31:31.438811547 +0000 UTC m=+1228.627042322" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.439062 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.458867 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536129 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536356 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55282f54-fc3a-418a-8eeb-3e4fa482b32b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536399 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536469 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536497 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536529 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g8n4\" (UniqueName: \"kubernetes.io/projected/55282f54-fc3a-418a-8eeb-3e4fa482b32b-kube-api-access-8g8n4\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536560 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.536598 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55282f54-fc3a-418a-8eeb-3e4fa482b32b-logs\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637769 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55282f54-fc3a-418a-8eeb-3e4fa482b32b-logs\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637839 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637869 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55282f54-fc3a-418a-8eeb-3e4fa482b32b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637909 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637937 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637963 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.637992 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g8n4\" (UniqueName: \"kubernetes.io/projected/55282f54-fc3a-418a-8eeb-3e4fa482b32b-kube-api-access-8g8n4\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.638024 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.638309 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.639641 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55282f54-fc3a-418a-8eeb-3e4fa482b32b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.639874 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55282f54-fc3a-418a-8eeb-3e4fa482b32b-logs\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.644375 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.644830 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.647031 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.647775 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55282f54-fc3a-418a-8eeb-3e4fa482b32b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.661249 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g8n4\" (UniqueName: \"kubernetes.io/projected/55282f54-fc3a-418a-8eeb-3e4fa482b32b-kube-api-access-8g8n4\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.682422 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"55282f54-fc3a-418a-8eeb-3e4fa482b32b\") " pod="openstack/glance-default-internal-api-0" Dec 04 06:31:31 crc kubenswrapper[4685]: I1204 06:31:31.944777 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.379890 4685 generic.go:334] "Generic (PLEG): container finished" podID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerID="86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f" exitCode=0 Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.379938 4685 generic.go:334] "Generic (PLEG): container finished" podID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerID="db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912" exitCode=2 Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.379957 4685 generic.go:334] "Generic (PLEG): container finished" podID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerID="60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812" exitCode=0 Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.379953 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerDied","Data":"86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f"} Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.379994 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerDied","Data":"db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912"} Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.380004 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerDied","Data":"60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812"} Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.381768 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c17edd9-e27f-4397-bfad-baeb4684038e","Type":"ContainerStarted","Data":"6dd756303385f7d0eb919670307426e2be90f4ce1c876b47c8d6214723d4161f"} Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.381822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c17edd9-e27f-4397-bfad-baeb4684038e","Type":"ContainerStarted","Data":"4c4b7b44fa5c36b3021c50aee7069d6ab4eb955a8ce5fc836d7c16481dd7c291"} Dec 04 06:31:32 crc kubenswrapper[4685]: I1204 06:31:32.583441 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 06:31:33 crc kubenswrapper[4685]: I1204 06:31:33.135608 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1b98dd-9b1f-474d-a6b0-7d0385db6a34" path="/var/lib/kubelet/pods/8d1b98dd-9b1f-474d-a6b0-7d0385db6a34/volumes" Dec 04 06:31:33 crc kubenswrapper[4685]: I1204 06:31:33.407189 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c17edd9-e27f-4397-bfad-baeb4684038e","Type":"ContainerStarted","Data":"3706065fac747f3a3306fd86a7e0bc939a03f5a3a81f8bc1e63c1702b9ffa158"} Dec 04 06:31:33 crc kubenswrapper[4685]: I1204 06:31:33.416134 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55282f54-fc3a-418a-8eeb-3e4fa482b32b","Type":"ContainerStarted","Data":"c4547fa30731238d1681f8953461223af643a9279d87599e70ee8f24a00c61cb"} Dec 04 06:31:33 crc kubenswrapper[4685]: I1204 06:31:33.416190 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55282f54-fc3a-418a-8eeb-3e4fa482b32b","Type":"ContainerStarted","Data":"711ddb14d430a2f0cf0031e78996cd9d1842dd1910f315cbe4759c7bbcde2fef"} Dec 04 06:31:33 crc kubenswrapper[4685]: I1204 06:31:33.434717 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.434697945 podStartE2EDuration="3.434697945s" podCreationTimestamp="2025-12-04 06:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:33.429098899 +0000 UTC m=+1230.617329704" watchObservedRunningTime="2025-12-04 06:31:33.434697945 +0000 UTC m=+1230.622928720" Dec 04 06:31:34 crc kubenswrapper[4685]: I1204 06:31:34.426110 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55282f54-fc3a-418a-8eeb-3e4fa482b32b","Type":"ContainerStarted","Data":"df0b738ec1ff9039474f9b23a96555d7929fe3b0b870b6c6b9b6cac90f333a8a"} Dec 04 06:31:34 crc kubenswrapper[4685]: I1204 06:31:34.480092 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.480062756 podStartE2EDuration="3.480062756s" podCreationTimestamp="2025-12-04 06:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:34.467258923 +0000 UTC m=+1231.655489708" watchObservedRunningTime="2025-12-04 06:31:34.480062756 +0000 UTC m=+1231.668293571" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.261900 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356146 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxkf4\" (UniqueName: \"kubernetes.io/projected/f41de5f3-3313-4d6b-80d8-3337a5aca42c-kube-api-access-bxkf4\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356261 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-scripts\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356298 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-log-httpd\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356368 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-combined-ca-bundle\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356455 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-sg-core-conf-yaml\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356491 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-config-data\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.356558 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-run-httpd\") pod \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\" (UID: \"f41de5f3-3313-4d6b-80d8-3337a5aca42c\") " Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.357303 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.358372 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.363558 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41de5f3-3313-4d6b-80d8-3337a5aca42c-kube-api-access-bxkf4" (OuterVolumeSpecName: "kube-api-access-bxkf4") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "kube-api-access-bxkf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.364612 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-scripts" (OuterVolumeSpecName: "scripts") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.393712 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.437293 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.458420 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxkf4\" (UniqueName: \"kubernetes.io/projected/f41de5f3-3313-4d6b-80d8-3337a5aca42c-kube-api-access-bxkf4\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.458450 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.458459 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.458467 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.458475 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.458485 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41de5f3-3313-4d6b-80d8-3337a5aca42c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.463919 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-config-data" (OuterVolumeSpecName: "config-data") pod "f41de5f3-3313-4d6b-80d8-3337a5aca42c" (UID: "f41de5f3-3313-4d6b-80d8-3337a5aca42c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.466685 4685 generic.go:334] "Generic (PLEG): container finished" podID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerID="efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88" exitCode=0 Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.466753 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerDied","Data":"efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88"} Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.466782 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41de5f3-3313-4d6b-80d8-3337a5aca42c","Type":"ContainerDied","Data":"99530c131eeb7f9e41609124f2463291c77584dfc6054188c17db31e1eab1f2b"} Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.466799 4685 scope.go:117] "RemoveContainer" containerID="86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.467127 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.469803 4685 generic.go:334] "Generic (PLEG): container finished" podID="18ec8793-df06-4f80-9ddc-148283651ef9" containerID="d3789629a7d33764b1db5b73e19ab06c1ea82860b04a1c59b836bba229fcfaf9" exitCode=0 Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.469838 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-chrlf" event={"ID":"18ec8793-df06-4f80-9ddc-148283651ef9","Type":"ContainerDied","Data":"d3789629a7d33764b1db5b73e19ab06c1ea82860b04a1c59b836bba229fcfaf9"} Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.504574 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.505818 4685 scope.go:117] "RemoveContainer" containerID="db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.511796 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.537548 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.538009 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-notification-agent" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538029 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-notification-agent" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.538050 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-central-agent" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538059 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-central-agent" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.538077 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="proxy-httpd" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538084 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="proxy-httpd" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.538107 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="sg-core" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538114 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="sg-core" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538325 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-central-agent" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538350 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="sg-core" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538367 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="ceilometer-notification-agent" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.538387 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" containerName="proxy-httpd" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.544484 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.547021 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.547326 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.548859 4685 scope.go:117] "RemoveContainer" containerID="60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.551012 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.560269 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41de5f3-3313-4d6b-80d8-3337a5aca42c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.582294 4685 scope.go:117] "RemoveContainer" containerID="efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.601977 4685 scope.go:117] "RemoveContainer" containerID="86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.602316 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f\": container with ID starting with 86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f not found: ID does not exist" containerID="86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.602368 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f"} err="failed to get container status \"86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f\": rpc error: code = NotFound desc = could not find container \"86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f\": container with ID starting with 86ec02fcbb6ebc77aa0f0cc01d7ced5710a752ffb8cce4e299945470659f0f1f not found: ID does not exist" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.602393 4685 scope.go:117] "RemoveContainer" containerID="db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.602789 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912\": container with ID starting with db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912 not found: ID does not exist" containerID="db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.602812 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912"} err="failed to get container status \"db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912\": rpc error: code = NotFound desc = could not find container \"db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912\": container with ID starting with db1ca1f5ff8809e3b65df4ddee58273f9f7b5dd7930aea3a0b9d59e217727912 not found: ID does not exist" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.602858 4685 scope.go:117] "RemoveContainer" containerID="60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.603294 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812\": container with ID starting with 60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812 not found: ID does not exist" containerID="60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.603335 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812"} err="failed to get container status \"60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812\": rpc error: code = NotFound desc = could not find container \"60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812\": container with ID starting with 60bd03cfa49445f3fcfe2a06600faefe833fb3437716a8c03af560e5c9d5e812 not found: ID does not exist" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.603360 4685 scope.go:117] "RemoveContainer" containerID="efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88" Dec 04 06:31:37 crc kubenswrapper[4685]: E1204 06:31:37.603988 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88\": container with ID starting with efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88 not found: ID does not exist" containerID="efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.604017 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88"} err="failed to get container status \"efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88\": rpc error: code = NotFound desc = could not find container \"efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88\": container with ID starting with efe522d7655d67feab170002787adcf7c4ffb5c4b6d2b8d05b7ac391e9f5cf88 not found: ID does not exist" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.661393 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.661459 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g29zt\" (UniqueName: \"kubernetes.io/projected/0d8dc337-7239-4721-a3f1-0f97e187865f-kube-api-access-g29zt\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.661483 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-config-data\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.661502 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.661824 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-scripts\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.662004 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-run-httpd\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.662048 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-log-httpd\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.765622 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-run-httpd\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.765700 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-log-httpd\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.765920 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.765979 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g29zt\" (UniqueName: \"kubernetes.io/projected/0d8dc337-7239-4721-a3f1-0f97e187865f-kube-api-access-g29zt\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.766026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-config-data\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.766060 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.766176 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-scripts\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.767957 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-run-httpd\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.768129 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-log-httpd\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.773342 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.775277 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-config-data\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.775459 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-scripts\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.776954 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.787631 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g29zt\" (UniqueName: \"kubernetes.io/projected/0d8dc337-7239-4721-a3f1-0f97e187865f-kube-api-access-g29zt\") pod \"ceilometer-0\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " pod="openstack/ceilometer-0" Dec 04 06:31:37 crc kubenswrapper[4685]: I1204 06:31:37.860193 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:31:38 crc kubenswrapper[4685]: I1204 06:31:38.359542 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:31:38 crc kubenswrapper[4685]: I1204 06:31:38.483011 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerStarted","Data":"b018329e41742ba013e398e8e5fca7ee14cdbca8b5fdb1220e601811dc428507"} Dec 04 06:31:38 crc kubenswrapper[4685]: I1204 06:31:38.875977 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.005588 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-config-data\") pod \"18ec8793-df06-4f80-9ddc-148283651ef9\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.005714 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-scripts\") pod \"18ec8793-df06-4f80-9ddc-148283651ef9\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.005778 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-combined-ca-bundle\") pod \"18ec8793-df06-4f80-9ddc-148283651ef9\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.005872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdzfd\" (UniqueName: \"kubernetes.io/projected/18ec8793-df06-4f80-9ddc-148283651ef9-kube-api-access-gdzfd\") pod \"18ec8793-df06-4f80-9ddc-148283651ef9\" (UID: \"18ec8793-df06-4f80-9ddc-148283651ef9\") " Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.013199 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-scripts" (OuterVolumeSpecName: "scripts") pod "18ec8793-df06-4f80-9ddc-148283651ef9" (UID: "18ec8793-df06-4f80-9ddc-148283651ef9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.013697 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ec8793-df06-4f80-9ddc-148283651ef9-kube-api-access-gdzfd" (OuterVolumeSpecName: "kube-api-access-gdzfd") pod "18ec8793-df06-4f80-9ddc-148283651ef9" (UID: "18ec8793-df06-4f80-9ddc-148283651ef9"). InnerVolumeSpecName "kube-api-access-gdzfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.045350 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18ec8793-df06-4f80-9ddc-148283651ef9" (UID: "18ec8793-df06-4f80-9ddc-148283651ef9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.046328 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-config-data" (OuterVolumeSpecName: "config-data") pod "18ec8793-df06-4f80-9ddc-148283651ef9" (UID: "18ec8793-df06-4f80-9ddc-148283651ef9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.108159 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.108686 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.108730 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdzfd\" (UniqueName: \"kubernetes.io/projected/18ec8793-df06-4f80-9ddc-148283651ef9-kube-api-access-gdzfd\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.108745 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ec8793-df06-4f80-9ddc-148283651ef9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.138817 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f41de5f3-3313-4d6b-80d8-3337a5aca42c" path="/var/lib/kubelet/pods/f41de5f3-3313-4d6b-80d8-3337a5aca42c/volumes" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.513913 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-chrlf" event={"ID":"18ec8793-df06-4f80-9ddc-148283651ef9","Type":"ContainerDied","Data":"9525921a559fadfac21db3f7b5c56f464f5dfa73b5860a75ca837ead3eaed8c2"} Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.514204 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9525921a559fadfac21db3f7b5c56f464f5dfa73b5860a75ca837ead3eaed8c2" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.514174 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-chrlf" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.586563 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 06:31:39 crc kubenswrapper[4685]: E1204 06:31:39.586993 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ec8793-df06-4f80-9ddc-148283651ef9" containerName="nova-cell0-conductor-db-sync" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.587009 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ec8793-df06-4f80-9ddc-148283651ef9" containerName="nova-cell0-conductor-db-sync" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.587201 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ec8793-df06-4f80-9ddc-148283651ef9" containerName="nova-cell0-conductor-db-sync" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.587808 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.589813 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.589960 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tpmkh" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.603108 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.619966 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44873bdc-8a05-43b8-897e-1f44d10563f7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.620029 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44873bdc-8a05-43b8-897e-1f44d10563f7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.620065 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn6tm\" (UniqueName: \"kubernetes.io/projected/44873bdc-8a05-43b8-897e-1f44d10563f7-kube-api-access-pn6tm\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.721402 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn6tm\" (UniqueName: \"kubernetes.io/projected/44873bdc-8a05-43b8-897e-1f44d10563f7-kube-api-access-pn6tm\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.721760 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44873bdc-8a05-43b8-897e-1f44d10563f7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.721913 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44873bdc-8a05-43b8-897e-1f44d10563f7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.726174 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44873bdc-8a05-43b8-897e-1f44d10563f7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.726258 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44873bdc-8a05-43b8-897e-1f44d10563f7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.738649 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn6tm\" (UniqueName: \"kubernetes.io/projected/44873bdc-8a05-43b8-897e-1f44d10563f7-kube-api-access-pn6tm\") pod \"nova-cell0-conductor-0\" (UID: \"44873bdc-8a05-43b8-897e-1f44d10563f7\") " pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:39 crc kubenswrapper[4685]: I1204 06:31:39.943855 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:40 crc kubenswrapper[4685]: I1204 06:31:40.526750 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerStarted","Data":"b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f"} Dec 04 06:31:40 crc kubenswrapper[4685]: I1204 06:31:40.776935 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 06:31:40 crc kubenswrapper[4685]: W1204 06:31:40.790086 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44873bdc_8a05_43b8_897e_1f44d10563f7.slice/crio-cd6eb86e9ef64c5e17f3ef23f4d1560e7137ec6a643a632b6f7c705ff4a2653e WatchSource:0}: Error finding container cd6eb86e9ef64c5e17f3ef23f4d1560e7137ec6a643a632b6f7c705ff4a2653e: Status 404 returned error can't find the container with id cd6eb86e9ef64c5e17f3ef23f4d1560e7137ec6a643a632b6f7c705ff4a2653e Dec 04 06:31:40 crc kubenswrapper[4685]: I1204 06:31:40.832321 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 06:31:40 crc kubenswrapper[4685]: I1204 06:31:40.833789 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 06:31:40 crc kubenswrapper[4685]: I1204 06:31:40.872004 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 06:31:40 crc kubenswrapper[4685]: I1204 06:31:40.884025 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.546912 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"44873bdc-8a05-43b8-897e-1f44d10563f7","Type":"ContainerStarted","Data":"6fe67ac2cdcb124ea9f04cbb0002bfca1b751dd8b4169cac5800c1acba4a3cc5"} Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.547276 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.547291 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"44873bdc-8a05-43b8-897e-1f44d10563f7","Type":"ContainerStarted","Data":"cd6eb86e9ef64c5e17f3ef23f4d1560e7137ec6a643a632b6f7c705ff4a2653e"} Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.550234 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerStarted","Data":"bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439"} Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.550513 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.550591 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.569854 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.569836462 podStartE2EDuration="2.569836462s" podCreationTimestamp="2025-12-04 06:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:41.566529299 +0000 UTC m=+1238.754760074" watchObservedRunningTime="2025-12-04 06:31:41.569836462 +0000 UTC m=+1238.758067237" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.945627 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.946037 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.979017 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:41 crc kubenswrapper[4685]: I1204 06:31:41.990305 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:42 crc kubenswrapper[4685]: I1204 06:31:42.568171 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerStarted","Data":"5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6"} Dec 04 06:31:42 crc kubenswrapper[4685]: I1204 06:31:42.569021 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:42 crc kubenswrapper[4685]: I1204 06:31:42.569203 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:43 crc kubenswrapper[4685]: I1204 06:31:43.423257 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 06:31:43 crc kubenswrapper[4685]: I1204 06:31:43.518295 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 06:31:43 crc kubenswrapper[4685]: I1204 06:31:43.576511 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerStarted","Data":"3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793"} Dec 04 06:31:43 crc kubenswrapper[4685]: I1204 06:31:43.603829 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.972419865 podStartE2EDuration="6.603806822s" podCreationTimestamp="2025-12-04 06:31:37 +0000 UTC" firstStartedPulling="2025-12-04 06:31:38.367599892 +0000 UTC m=+1235.555830677" lastFinishedPulling="2025-12-04 06:31:42.998986859 +0000 UTC m=+1240.187217634" observedRunningTime="2025-12-04 06:31:43.603041067 +0000 UTC m=+1240.791271852" watchObservedRunningTime="2025-12-04 06:31:43.603806822 +0000 UTC m=+1240.792037597" Dec 04 06:31:44 crc kubenswrapper[4685]: I1204 06:31:44.589471 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:31:44 crc kubenswrapper[4685]: I1204 06:31:44.766190 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:44 crc kubenswrapper[4685]: I1204 06:31:44.766347 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 06:31:44 crc kubenswrapper[4685]: I1204 06:31:44.771777 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 06:31:49 crc kubenswrapper[4685]: I1204 06:31:49.995216 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.512790 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-pp2fn"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.514534 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.516282 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.517520 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.526397 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pp2fn"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.586041 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.586108 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shsxp\" (UniqueName: \"kubernetes.io/projected/8073c264-8fb4-4316-8b24-5de5ca8238be-kube-api-access-shsxp\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.586173 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-scripts\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.586226 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-config-data\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.687835 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.687890 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shsxp\" (UniqueName: \"kubernetes.io/projected/8073c264-8fb4-4316-8b24-5de5ca8238be-kube-api-access-shsxp\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.687954 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-scripts\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.688315 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-config-data\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.695035 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-scripts\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.695154 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-config-data\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.706876 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.707554 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shsxp\" (UniqueName: \"kubernetes.io/projected/8073c264-8fb4-4316-8b24-5de5ca8238be-kube-api-access-shsxp\") pod \"nova-cell0-cell-mapping-pp2fn\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.733307 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.734852 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.742342 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.746570 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.774748 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.775985 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.781831 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.789975 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd5xq\" (UniqueName: \"kubernetes.io/projected/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-kube-api-access-gd5xq\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.790082 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.790107 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-config-data\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.790386 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.837271 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.882528 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.891768 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.891819 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mbsw\" (UniqueName: \"kubernetes.io/projected/70d66855-2309-4bd4-9929-794bd475dd3f-kube-api-access-2mbsw\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.891846 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-config-data\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.891880 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.891914 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.891966 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd5xq\" (UniqueName: \"kubernetes.io/projected/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-kube-api-access-gd5xq\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.893978 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.899775 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.900753 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.933429 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-config-data\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.910514 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.941973 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.943545 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.951235 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.957022 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.959052 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd5xq\" (UniqueName: \"kubernetes.io/projected/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-kube-api-access-gd5xq\") pod \"nova-scheduler-0\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " pod="openstack/nova-scheduler-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994710 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mbsw\" (UniqueName: \"kubernetes.io/projected/70d66855-2309-4bd4-9929-794bd475dd3f-kube-api-access-2mbsw\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994773 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-config-data\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994799 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994817 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9qmv\" (UniqueName: \"kubernetes.io/projected/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-kube-api-access-k9qmv\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994844 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eec7f81-a673-43e1-841b-28c650bba0a6-logs\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994860 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmtbm\" (UniqueName: \"kubernetes.io/projected/4eec7f81-a673-43e1-841b-28c650bba0a6-kube-api-access-wmtbm\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994885 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994903 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-config-data\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994926 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-logs\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.994979 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:50 crc kubenswrapper[4685]: I1204 06:31:50.995007 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.003676 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.005020 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.037914 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mbsw\" (UniqueName: \"kubernetes.io/projected/70d66855-2309-4bd4-9929-794bd475dd3f-kube-api-access-2mbsw\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.052946 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-psqjs"] Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.054828 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.058821 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-psqjs"] Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097074 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-config\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097121 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097153 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ltwv\" (UniqueName: \"kubernetes.io/projected/15c9f915-a417-4034-9d34-2a2b1474a104-kube-api-access-6ltwv\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097198 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-config-data\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097220 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9qmv\" (UniqueName: \"kubernetes.io/projected/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-kube-api-access-k9qmv\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097242 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eec7f81-a673-43e1-841b-28c650bba0a6-logs\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097256 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmtbm\" (UniqueName: \"kubernetes.io/projected/4eec7f81-a673-43e1-841b-28c650bba0a6-kube-api-access-wmtbm\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097274 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097298 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-config-data\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097318 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-logs\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097357 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097381 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097398 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-svc\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.097431 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.103003 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eec7f81-a673-43e1-841b-28c650bba0a6-logs\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.103366 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-logs\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.105304 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.105490 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-config-data\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.106436 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-config-data\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.106680 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.121449 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmtbm\" (UniqueName: \"kubernetes.io/projected/4eec7f81-a673-43e1-841b-28c650bba0a6-kube-api-access-wmtbm\") pod \"nova-metadata-0\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.125332 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9qmv\" (UniqueName: \"kubernetes.io/projected/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-kube-api-access-k9qmv\") pod \"nova-api-0\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.139324 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.150417 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.199135 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.200258 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ltwv\" (UniqueName: \"kubernetes.io/projected/15c9f915-a417-4034-9d34-2a2b1474a104-kube-api-access-6ltwv\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.200187 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.200786 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.201541 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.201681 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.202256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.202284 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-svc\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.202376 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-config\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.203216 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-config\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.203465 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-svc\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.220907 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ltwv\" (UniqueName: \"kubernetes.io/projected/15c9f915-a417-4034-9d34-2a2b1474a104-kube-api-access-6ltwv\") pod \"dnsmasq-dns-bccf8f775-psqjs\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.337266 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.354327 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.371070 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.522361 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pp2fn"] Dec 04 06:31:51 crc kubenswrapper[4685]: W1204 06:31:51.531943 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8073c264_8fb4_4316_8b24_5de5ca8238be.slice/crio-41e2183a392f44dee2de8df629642cf5993b02b60530d55bb467301ab608e9f5 WatchSource:0}: Error finding container 41e2183a392f44dee2de8df629642cf5993b02b60530d55bb467301ab608e9f5: Status 404 returned error can't find the container with id 41e2183a392f44dee2de8df629642cf5993b02b60530d55bb467301ab608e9f5 Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.634393 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.689489 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pp2fn" event={"ID":"8073c264-8fb4-4316-8b24-5de5ca8238be","Type":"ContainerStarted","Data":"41e2183a392f44dee2de8df629642cf5993b02b60530d55bb467301ab608e9f5"} Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.692521 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c","Type":"ContainerStarted","Data":"35b6473f84ae1a165725df72bf5c71fef6293e3054752de5303acd8617e11fc5"} Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.709909 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.750731 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zc8bc"] Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.755143 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.760202 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.760354 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.775235 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zc8bc"] Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.867783 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:51 crc kubenswrapper[4685]: W1204 06:31:51.874952 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eec7f81_a673_43e1_841b_28c650bba0a6.slice/crio-d5abd7668b109054d9fae5e5bf7b942befb2fe60086ebd972302f94b294e49c0 WatchSource:0}: Error finding container d5abd7668b109054d9fae5e5bf7b942befb2fe60086ebd972302f94b294e49c0: Status 404 returned error can't find the container with id d5abd7668b109054d9fae5e5bf7b942befb2fe60086ebd972302f94b294e49c0 Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.923674 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-scripts\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.923746 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz6z8\" (UniqueName: \"kubernetes.io/projected/b37a166d-763c-4492-bcc8-fef7326f7d82-kube-api-access-dz6z8\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.923770 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-config-data\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.923958 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:51 crc kubenswrapper[4685]: I1204 06:31:51.963080 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.026234 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.026368 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-scripts\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.026428 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz6z8\" (UniqueName: \"kubernetes.io/projected/b37a166d-763c-4492-bcc8-fef7326f7d82-kube-api-access-dz6z8\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.026447 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-config-data\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.031089 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.034103 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-config-data\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.034785 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-scripts\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.041934 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz6z8\" (UniqueName: \"kubernetes.io/projected/b37a166d-763c-4492-bcc8-fef7326f7d82-kube-api-access-dz6z8\") pod \"nova-cell1-conductor-db-sync-zc8bc\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.103576 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-psqjs"] Dec 04 06:31:52 crc kubenswrapper[4685]: W1204 06:31:52.109076 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15c9f915_a417_4034_9d34_2a2b1474a104.slice/crio-e72f144f62838fef8ccde47e8ef7e5ba0ec372cf7f633a98b3f4e4083ae87d25 WatchSource:0}: Error finding container e72f144f62838fef8ccde47e8ef7e5ba0ec372cf7f633a98b3f4e4083ae87d25: Status 404 returned error can't find the container with id e72f144f62838fef8ccde47e8ef7e5ba0ec372cf7f633a98b3f4e4083ae87d25 Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.110590 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.600869 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zc8bc"] Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.711811 4685 generic.go:334] "Generic (PLEG): container finished" podID="15c9f915-a417-4034-9d34-2a2b1474a104" containerID="cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819" exitCode=0 Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.711900 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" event={"ID":"15c9f915-a417-4034-9d34-2a2b1474a104","Type":"ContainerDied","Data":"cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819"} Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.711941 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" event={"ID":"15c9f915-a417-4034-9d34-2a2b1474a104","Type":"ContainerStarted","Data":"e72f144f62838fef8ccde47e8ef7e5ba0ec372cf7f633a98b3f4e4083ae87d25"} Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.715021 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4eec7f81-a673-43e1-841b-28c650bba0a6","Type":"ContainerStarted","Data":"d5abd7668b109054d9fae5e5bf7b942befb2fe60086ebd972302f94b294e49c0"} Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.721368 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pp2fn" event={"ID":"8073c264-8fb4-4316-8b24-5de5ca8238be","Type":"ContainerStarted","Data":"b857ed3a1b6e515b527451457d9dd77d94d20b4ced96da049ad39e13b342e285"} Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.724345 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6f4e7cf1-5021-40ad-9c87-2f0264b339fa","Type":"ContainerStarted","Data":"9daa586a48b3ce4485843d33d837ffa1bb0b945743667c05b31ae173539b5066"} Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.732018 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d66855-2309-4bd4-9929-794bd475dd3f","Type":"ContainerStarted","Data":"2b0426d0873a97db989a91ec526ece6811981fb645dd6b70c6a80764d790c0fd"} Dec 04 06:31:52 crc kubenswrapper[4685]: I1204 06:31:52.759661 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-pp2fn" podStartSLOduration=2.759642586 podStartE2EDuration="2.759642586s" podCreationTimestamp="2025-12-04 06:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:52.74708132 +0000 UTC m=+1249.935312095" watchObservedRunningTime="2025-12-04 06:31:52.759642586 +0000 UTC m=+1249.947873371" Dec 04 06:31:53 crc kubenswrapper[4685]: I1204 06:31:53.745007 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" event={"ID":"b37a166d-763c-4492-bcc8-fef7326f7d82","Type":"ContainerStarted","Data":"e5525345b481ea096d437ba1a91cf0ae0663191b1e5147c4ce57da2953e9f7be"} Dec 04 06:31:54 crc kubenswrapper[4685]: I1204 06:31:54.527437 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:54 crc kubenswrapper[4685]: I1204 06:31:54.537273 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.765505 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" event={"ID":"15c9f915-a417-4034-9d34-2a2b1474a104","Type":"ContainerStarted","Data":"d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.766025 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.768700 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4eec7f81-a673-43e1-841b-28c650bba0a6","Type":"ContainerStarted","Data":"037bcd5d9c5e1a269be02ec8f9a987e72891d759a570d0b8235167400f87a243"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.768731 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4eec7f81-a673-43e1-841b-28c650bba0a6","Type":"ContainerStarted","Data":"bc4b1131ba7c64eacad612b27d0cde155fe0fadffa3bf7d74fe94635678af5d1"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.768919 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-log" containerID="cri-o://bc4b1131ba7c64eacad612b27d0cde155fe0fadffa3bf7d74fe94635678af5d1" gracePeriod=30 Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.768965 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-metadata" containerID="cri-o://037bcd5d9c5e1a269be02ec8f9a987e72891d759a570d0b8235167400f87a243" gracePeriod=30 Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.770871 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6f4e7cf1-5021-40ad-9c87-2f0264b339fa","Type":"ContainerStarted","Data":"800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.770971 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6f4e7cf1-5021-40ad-9c87-2f0264b339fa","Type":"ContainerStarted","Data":"c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.781068 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d66855-2309-4bd4-9929-794bd475dd3f","Type":"ContainerStarted","Data":"1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.781096 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="70d66855-2309-4bd4-9929-794bd475dd3f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03" gracePeriod=30 Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.783023 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c","Type":"ContainerStarted","Data":"190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.789832 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" podStartSLOduration=5.789814234 podStartE2EDuration="5.789814234s" podCreationTimestamp="2025-12-04 06:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:55.786633354 +0000 UTC m=+1252.974864129" watchObservedRunningTime="2025-12-04 06:31:55.789814234 +0000 UTC m=+1252.978044999" Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.797024 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" event={"ID":"b37a166d-763c-4492-bcc8-fef7326f7d82","Type":"ContainerStarted","Data":"cb57c41e0d01aacf3bce298f5de0330ebb3d9388841c18402625385e47f3efd8"} Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.811304 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.024766706 podStartE2EDuration="5.811283982s" podCreationTimestamp="2025-12-04 06:31:50 +0000 UTC" firstStartedPulling="2025-12-04 06:31:51.976556666 +0000 UTC m=+1249.164787441" lastFinishedPulling="2025-12-04 06:31:54.763073942 +0000 UTC m=+1251.951304717" observedRunningTime="2025-12-04 06:31:55.806759989 +0000 UTC m=+1252.994990784" watchObservedRunningTime="2025-12-04 06:31:55.811283982 +0000 UTC m=+1252.999514757" Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.837145 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.777509061 podStartE2EDuration="5.837128458s" podCreationTimestamp="2025-12-04 06:31:50 +0000 UTC" firstStartedPulling="2025-12-04 06:31:51.696874357 +0000 UTC m=+1248.885105132" lastFinishedPulling="2025-12-04 06:31:54.756493744 +0000 UTC m=+1251.944724529" observedRunningTime="2025-12-04 06:31:55.828579568 +0000 UTC m=+1253.016810343" watchObservedRunningTime="2025-12-04 06:31:55.837128458 +0000 UTC m=+1253.025359233" Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.850578 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.974163239 podStartE2EDuration="5.850561302s" podCreationTimestamp="2025-12-04 06:31:50 +0000 UTC" firstStartedPulling="2025-12-04 06:31:51.878921154 +0000 UTC m=+1249.067151929" lastFinishedPulling="2025-12-04 06:31:54.755319217 +0000 UTC m=+1251.943549992" observedRunningTime="2025-12-04 06:31:55.84574286 +0000 UTC m=+1253.033973635" watchObservedRunningTime="2025-12-04 06:31:55.850561302 +0000 UTC m=+1253.038792077" Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.874946 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.756779837 podStartE2EDuration="5.874924621s" podCreationTimestamp="2025-12-04 06:31:50 +0000 UTC" firstStartedPulling="2025-12-04 06:31:51.638700841 +0000 UTC m=+1248.826931616" lastFinishedPulling="2025-12-04 06:31:54.756845625 +0000 UTC m=+1251.945076400" observedRunningTime="2025-12-04 06:31:55.866799865 +0000 UTC m=+1253.055030650" watchObservedRunningTime="2025-12-04 06:31:55.874924621 +0000 UTC m=+1253.063155396" Dec 04 06:31:55 crc kubenswrapper[4685]: I1204 06:31:55.889496 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" podStartSLOduration=4.88948013 podStartE2EDuration="4.88948013s" podCreationTimestamp="2025-12-04 06:31:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:55.88787217 +0000 UTC m=+1253.076102945" watchObservedRunningTime="2025-12-04 06:31:55.88948013 +0000 UTC m=+1253.077710905" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.142217 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.151537 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.338712 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.338757 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.808547 4685 generic.go:334] "Generic (PLEG): container finished" podID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerID="037bcd5d9c5e1a269be02ec8f9a987e72891d759a570d0b8235167400f87a243" exitCode=0 Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.808849 4685 generic.go:334] "Generic (PLEG): container finished" podID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerID="bc4b1131ba7c64eacad612b27d0cde155fe0fadffa3bf7d74fe94635678af5d1" exitCode=143 Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.808639 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4eec7f81-a673-43e1-841b-28c650bba0a6","Type":"ContainerDied","Data":"037bcd5d9c5e1a269be02ec8f9a987e72891d759a570d0b8235167400f87a243"} Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.809023 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4eec7f81-a673-43e1-841b-28c650bba0a6","Type":"ContainerDied","Data":"bc4b1131ba7c64eacad612b27d0cde155fe0fadffa3bf7d74fe94635678af5d1"} Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.809043 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4eec7f81-a673-43e1-841b-28c650bba0a6","Type":"ContainerDied","Data":"d5abd7668b109054d9fae5e5bf7b942befb2fe60086ebd972302f94b294e49c0"} Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.809056 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5abd7668b109054d9fae5e5bf7b942befb2fe60086ebd972302f94b294e49c0" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.820747 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.952680 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eec7f81-a673-43e1-841b-28c650bba0a6-logs\") pod \"4eec7f81-a673-43e1-841b-28c650bba0a6\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.952801 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-config-data\") pod \"4eec7f81-a673-43e1-841b-28c650bba0a6\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.952916 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-combined-ca-bundle\") pod \"4eec7f81-a673-43e1-841b-28c650bba0a6\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.952946 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmtbm\" (UniqueName: \"kubernetes.io/projected/4eec7f81-a673-43e1-841b-28c650bba0a6-kube-api-access-wmtbm\") pod \"4eec7f81-a673-43e1-841b-28c650bba0a6\" (UID: \"4eec7f81-a673-43e1-841b-28c650bba0a6\") " Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.954293 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eec7f81-a673-43e1-841b-28c650bba0a6-logs" (OuterVolumeSpecName: "logs") pod "4eec7f81-a673-43e1-841b-28c650bba0a6" (UID: "4eec7f81-a673-43e1-841b-28c650bba0a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.974682 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eec7f81-a673-43e1-841b-28c650bba0a6-kube-api-access-wmtbm" (OuterVolumeSpecName: "kube-api-access-wmtbm") pod "4eec7f81-a673-43e1-841b-28c650bba0a6" (UID: "4eec7f81-a673-43e1-841b-28c650bba0a6"). InnerVolumeSpecName "kube-api-access-wmtbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:31:56 crc kubenswrapper[4685]: I1204 06:31:56.990874 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4eec7f81-a673-43e1-841b-28c650bba0a6" (UID: "4eec7f81-a673-43e1-841b-28c650bba0a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.008597 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-config-data" (OuterVolumeSpecName: "config-data") pod "4eec7f81-a673-43e1-841b-28c650bba0a6" (UID: "4eec7f81-a673-43e1-841b-28c650bba0a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.055140 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.055178 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec7f81-a673-43e1-841b-28c650bba0a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.055193 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmtbm\" (UniqueName: \"kubernetes.io/projected/4eec7f81-a673-43e1-841b-28c650bba0a6-kube-api-access-wmtbm\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.055207 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eec7f81-a673-43e1-841b-28c650bba0a6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.819203 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.851038 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.864186 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.886604 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:57 crc kubenswrapper[4685]: E1204 06:31:57.887106 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-log" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.887128 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-log" Dec 04 06:31:57 crc kubenswrapper[4685]: E1204 06:31:57.887149 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-metadata" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.887158 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-metadata" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.887390 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-metadata" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.887430 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" containerName="nova-metadata-log" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.888419 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.891333 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.891851 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.918271 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.973248 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-config-data\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.973344 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.973508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp89v\" (UniqueName: \"kubernetes.io/projected/c870b34c-9387-4fbe-b26a-a487d5bd022a-kube-api-access-tp89v\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.973541 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c870b34c-9387-4fbe-b26a-a487d5bd022a-logs\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:57 crc kubenswrapper[4685]: I1204 06:31:57.973635 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.075662 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.075731 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-config-data\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.075794 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.075894 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp89v\" (UniqueName: \"kubernetes.io/projected/c870b34c-9387-4fbe-b26a-a487d5bd022a-kube-api-access-tp89v\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.075922 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c870b34c-9387-4fbe-b26a-a487d5bd022a-logs\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.077144 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c870b34c-9387-4fbe-b26a-a487d5bd022a-logs\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.080593 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-config-data\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.080770 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.081254 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.095060 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp89v\" (UniqueName: \"kubernetes.io/projected/c870b34c-9387-4fbe-b26a-a487d5bd022a-kube-api-access-tp89v\") pod \"nova-metadata-0\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.220851 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:31:58 crc kubenswrapper[4685]: W1204 06:31:58.710813 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc870b34c_9387_4fbe_b26a_a487d5bd022a.slice/crio-d042aa7dd236f9e1a66d51275431d2db699084c3abe0f22cd8e02b7f05b23c93 WatchSource:0}: Error finding container d042aa7dd236f9e1a66d51275431d2db699084c3abe0f22cd8e02b7f05b23c93: Status 404 returned error can't find the container with id d042aa7dd236f9e1a66d51275431d2db699084c3abe0f22cd8e02b7f05b23c93 Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.722895 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:31:58 crc kubenswrapper[4685]: I1204 06:31:58.827292 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c870b34c-9387-4fbe-b26a-a487d5bd022a","Type":"ContainerStarted","Data":"d042aa7dd236f9e1a66d51275431d2db699084c3abe0f22cd8e02b7f05b23c93"} Dec 04 06:31:59 crc kubenswrapper[4685]: I1204 06:31:59.137015 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eec7f81-a673-43e1-841b-28c650bba0a6" path="/var/lib/kubelet/pods/4eec7f81-a673-43e1-841b-28c650bba0a6/volumes" Dec 04 06:31:59 crc kubenswrapper[4685]: I1204 06:31:59.846033 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c870b34c-9387-4fbe-b26a-a487d5bd022a","Type":"ContainerStarted","Data":"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a"} Dec 04 06:31:59 crc kubenswrapper[4685]: I1204 06:31:59.846489 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c870b34c-9387-4fbe-b26a-a487d5bd022a","Type":"ContainerStarted","Data":"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a"} Dec 04 06:31:59 crc kubenswrapper[4685]: I1204 06:31:59.848729 4685 generic.go:334] "Generic (PLEG): container finished" podID="8073c264-8fb4-4316-8b24-5de5ca8238be" containerID="b857ed3a1b6e515b527451457d9dd77d94d20b4ced96da049ad39e13b342e285" exitCode=0 Dec 04 06:31:59 crc kubenswrapper[4685]: I1204 06:31:59.848779 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pp2fn" event={"ID":"8073c264-8fb4-4316-8b24-5de5ca8238be","Type":"ContainerDied","Data":"b857ed3a1b6e515b527451457d9dd77d94d20b4ced96da049ad39e13b342e285"} Dec 04 06:31:59 crc kubenswrapper[4685]: I1204 06:31:59.882098 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.88206653 podStartE2EDuration="2.88206653s" podCreationTimestamp="2025-12-04 06:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:31:59.87222027 +0000 UTC m=+1257.060451045" watchObservedRunningTime="2025-12-04 06:31:59.88206653 +0000 UTC m=+1257.070297305" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.140743 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.172088 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.308183 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.355467 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.355525 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.372859 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.432622 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-2g4gj"] Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.432879 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerName="dnsmasq-dns" containerID="cri-o://365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0" gracePeriod=10 Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.449826 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-scripts\") pod \"8073c264-8fb4-4316-8b24-5de5ca8238be\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.449988 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-combined-ca-bundle\") pod \"8073c264-8fb4-4316-8b24-5de5ca8238be\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.450142 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shsxp\" (UniqueName: \"kubernetes.io/projected/8073c264-8fb4-4316-8b24-5de5ca8238be-kube-api-access-shsxp\") pod \"8073c264-8fb4-4316-8b24-5de5ca8238be\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.450197 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-config-data\") pod \"8073c264-8fb4-4316-8b24-5de5ca8238be\" (UID: \"8073c264-8fb4-4316-8b24-5de5ca8238be\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.459723 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-scripts" (OuterVolumeSpecName: "scripts") pod "8073c264-8fb4-4316-8b24-5de5ca8238be" (UID: "8073c264-8fb4-4316-8b24-5de5ca8238be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.459862 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8073c264-8fb4-4316-8b24-5de5ca8238be-kube-api-access-shsxp" (OuterVolumeSpecName: "kube-api-access-shsxp") pod "8073c264-8fb4-4316-8b24-5de5ca8238be" (UID: "8073c264-8fb4-4316-8b24-5de5ca8238be"). InnerVolumeSpecName "kube-api-access-shsxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.481647 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8073c264-8fb4-4316-8b24-5de5ca8238be" (UID: "8073c264-8fb4-4316-8b24-5de5ca8238be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.514705 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-config-data" (OuterVolumeSpecName: "config-data") pod "8073c264-8fb4-4316-8b24-5de5ca8238be" (UID: "8073c264-8fb4-4316-8b24-5de5ca8238be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.552145 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.552183 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.552192 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8073c264-8fb4-4316-8b24-5de5ca8238be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.552203 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shsxp\" (UniqueName: \"kubernetes.io/projected/8073c264-8fb4-4316-8b24-5de5ca8238be-kube-api-access-shsxp\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.847269 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.883369 4685 generic.go:334] "Generic (PLEG): container finished" podID="b37a166d-763c-4492-bcc8-fef7326f7d82" containerID="cb57c41e0d01aacf3bce298f5de0330ebb3d9388841c18402625385e47f3efd8" exitCode=0 Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.883488 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" event={"ID":"b37a166d-763c-4492-bcc8-fef7326f7d82","Type":"ContainerDied","Data":"cb57c41e0d01aacf3bce298f5de0330ebb3d9388841c18402625385e47f3efd8"} Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.885786 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pp2fn" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.885829 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pp2fn" event={"ID":"8073c264-8fb4-4316-8b24-5de5ca8238be","Type":"ContainerDied","Data":"41e2183a392f44dee2de8df629642cf5993b02b60530d55bb467301ab608e9f5"} Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.885874 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41e2183a392f44dee2de8df629642cf5993b02b60530d55bb467301ab608e9f5" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.892052 4685 generic.go:334] "Generic (PLEG): container finished" podID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerID="365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0" exitCode=0 Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.892229 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" event={"ID":"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d","Type":"ContainerDied","Data":"365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0"} Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.892297 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" event={"ID":"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d","Type":"ContainerDied","Data":"c5beb3197900020d80843b999bb22944bff0a80cfd7e7c4ac715dc9f1597cb70"} Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.892326 4685 scope.go:117] "RemoveContainer" containerID="365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.892693 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-2g4gj" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.931668 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.931736 4685 scope.go:117] "RemoveContainer" containerID="95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.959471 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-config\") pod \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.959596 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-svc\") pod \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.959649 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k6tz\" (UniqueName: \"kubernetes.io/projected/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-kube-api-access-2k6tz\") pod \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.959723 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-swift-storage-0\") pod \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.959789 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-sb\") pod \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.959872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-nb\") pod \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\" (UID: \"88c9de0e-75a3-43d1-bbfd-ca70847e1b6d\") " Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.964982 4685 scope.go:117] "RemoveContainer" containerID="365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0" Dec 04 06:32:01 crc kubenswrapper[4685]: E1204 06:32:01.965642 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0\": container with ID starting with 365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0 not found: ID does not exist" containerID="365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.965717 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0"} err="failed to get container status \"365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0\": rpc error: code = NotFound desc = could not find container \"365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0\": container with ID starting with 365d7dffc9085d3052784611db214f365737f0112c94efe7130a0fd842d652e0 not found: ID does not exist" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.965751 4685 scope.go:117] "RemoveContainer" containerID="95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0" Dec 04 06:32:01 crc kubenswrapper[4685]: E1204 06:32:01.966108 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0\": container with ID starting with 95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0 not found: ID does not exist" containerID="95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.966128 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0"} err="failed to get container status \"95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0\": rpc error: code = NotFound desc = could not find container \"95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0\": container with ID starting with 95e8df6f76d16326f9ff8c8bd05c8083ae3639df8d2a0172ba87da803a29f7f0 not found: ID does not exist" Dec 04 06:32:01 crc kubenswrapper[4685]: I1204 06:32:01.974241 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-kube-api-access-2k6tz" (OuterVolumeSpecName: "kube-api-access-2k6tz") pod "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" (UID: "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d"). InnerVolumeSpecName "kube-api-access-2k6tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.014685 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-config" (OuterVolumeSpecName: "config") pod "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" (UID: "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.019439 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" (UID: "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.021160 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" (UID: "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.027263 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" (UID: "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.030654 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" (UID: "88c9de0e-75a3-43d1-bbfd-ca70847e1b6d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.061940 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.061971 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.061980 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.061989 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k6tz\" (UniqueName: \"kubernetes.io/projected/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-kube-api-access-2k6tz\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.062000 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.062008 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.070224 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.070492 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-log" containerID="cri-o://c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b" gracePeriod=30 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.070837 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-api" containerID="cri-o://800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f" gracePeriod=30 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.078688 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": EOF" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.079215 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": EOF" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.104227 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.104494 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-log" containerID="cri-o://7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a" gracePeriod=30 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.104577 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-metadata" containerID="cri-o://1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a" gracePeriod=30 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.226143 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-2g4gj"] Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.237736 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-2g4gj"] Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.372116 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.761189 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.877127 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c870b34c-9387-4fbe-b26a-a487d5bd022a-logs\") pod \"c870b34c-9387-4fbe-b26a-a487d5bd022a\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.877199 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp89v\" (UniqueName: \"kubernetes.io/projected/c870b34c-9387-4fbe-b26a-a487d5bd022a-kube-api-access-tp89v\") pod \"c870b34c-9387-4fbe-b26a-a487d5bd022a\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.877675 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c870b34c-9387-4fbe-b26a-a487d5bd022a-logs" (OuterVolumeSpecName: "logs") pod "c870b34c-9387-4fbe-b26a-a487d5bd022a" (UID: "c870b34c-9387-4fbe-b26a-a487d5bd022a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.878233 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-combined-ca-bundle\") pod \"c870b34c-9387-4fbe-b26a-a487d5bd022a\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.878311 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-config-data\") pod \"c870b34c-9387-4fbe-b26a-a487d5bd022a\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.878337 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-nova-metadata-tls-certs\") pod \"c870b34c-9387-4fbe-b26a-a487d5bd022a\" (UID: \"c870b34c-9387-4fbe-b26a-a487d5bd022a\") " Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.878981 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c870b34c-9387-4fbe-b26a-a487d5bd022a-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.880921 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c870b34c-9387-4fbe-b26a-a487d5bd022a-kube-api-access-tp89v" (OuterVolumeSpecName: "kube-api-access-tp89v") pod "c870b34c-9387-4fbe-b26a-a487d5bd022a" (UID: "c870b34c-9387-4fbe-b26a-a487d5bd022a"). InnerVolumeSpecName "kube-api-access-tp89v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.903512 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-config-data" (OuterVolumeSpecName: "config-data") pod "c870b34c-9387-4fbe-b26a-a487d5bd022a" (UID: "c870b34c-9387-4fbe-b26a-a487d5bd022a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.910335 4685 generic.go:334] "Generic (PLEG): container finished" podID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerID="c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b" exitCode=143 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.910439 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6f4e7cf1-5021-40ad-9c87-2f0264b339fa","Type":"ContainerDied","Data":"c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b"} Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.912649 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c870b34c-9387-4fbe-b26a-a487d5bd022a" (UID: "c870b34c-9387-4fbe-b26a-a487d5bd022a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.914508 4685 generic.go:334] "Generic (PLEG): container finished" podID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerID="1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a" exitCode=0 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.914538 4685 generic.go:334] "Generic (PLEG): container finished" podID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerID="7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a" exitCode=143 Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.915378 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.915870 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c870b34c-9387-4fbe-b26a-a487d5bd022a","Type":"ContainerDied","Data":"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a"} Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.915903 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c870b34c-9387-4fbe-b26a-a487d5bd022a","Type":"ContainerDied","Data":"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a"} Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.915918 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c870b34c-9387-4fbe-b26a-a487d5bd022a","Type":"ContainerDied","Data":"d042aa7dd236f9e1a66d51275431d2db699084c3abe0f22cd8e02b7f05b23c93"} Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.915936 4685 scope.go:117] "RemoveContainer" containerID="1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.941815 4685 scope.go:117] "RemoveContainer" containerID="7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.942925 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c870b34c-9387-4fbe-b26a-a487d5bd022a" (UID: "c870b34c-9387-4fbe-b26a-a487d5bd022a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.973995 4685 scope.go:117] "RemoveContainer" containerID="1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a" Dec 04 06:32:02 crc kubenswrapper[4685]: E1204 06:32:02.975898 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a\": container with ID starting with 1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a not found: ID does not exist" containerID="1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.976529 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a"} err="failed to get container status \"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a\": rpc error: code = NotFound desc = could not find container \"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a\": container with ID starting with 1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a not found: ID does not exist" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.976554 4685 scope.go:117] "RemoveContainer" containerID="7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a" Dec 04 06:32:02 crc kubenswrapper[4685]: E1204 06:32:02.977350 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a\": container with ID starting with 7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a not found: ID does not exist" containerID="7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.977394 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a"} err="failed to get container status \"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a\": rpc error: code = NotFound desc = could not find container \"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a\": container with ID starting with 7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a not found: ID does not exist" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.977592 4685 scope.go:117] "RemoveContainer" containerID="1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.977818 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a"} err="failed to get container status \"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a\": rpc error: code = NotFound desc = could not find container \"1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a\": container with ID starting with 1c2020ef92fde78a8a632f53da9222dc1c9fff313951b71765cb55e031479d0a not found: ID does not exist" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.977833 4685 scope.go:117] "RemoveContainer" containerID="7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.978017 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a"} err="failed to get container status \"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a\": rpc error: code = NotFound desc = could not find container \"7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a\": container with ID starting with 7f42c94ade75ba55b19f2797a697aa49a608c5f6c9dc4bc2a066938f1406034a not found: ID does not exist" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.980237 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp89v\" (UniqueName: \"kubernetes.io/projected/c870b34c-9387-4fbe-b26a-a487d5bd022a-kube-api-access-tp89v\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.980258 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.980268 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:02 crc kubenswrapper[4685]: I1204 06:32:02.980278 4685 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c870b34c-9387-4fbe-b26a-a487d5bd022a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.161366 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" path="/var/lib/kubelet/pods/88c9de0e-75a3-43d1-bbfd-ca70847e1b6d/volumes" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.324080 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.342448 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.349373 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.370793 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:03 crc kubenswrapper[4685]: E1204 06:32:03.371281 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerName="init" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.374767 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerName="init" Dec 04 06:32:03 crc kubenswrapper[4685]: E1204 06:32:03.374819 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b37a166d-763c-4492-bcc8-fef7326f7d82" containerName="nova-cell1-conductor-db-sync" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.374829 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b37a166d-763c-4492-bcc8-fef7326f7d82" containerName="nova-cell1-conductor-db-sync" Dec 04 06:32:03 crc kubenswrapper[4685]: E1204 06:32:03.374858 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-log" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.374867 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-log" Dec 04 06:32:03 crc kubenswrapper[4685]: E1204 06:32:03.374891 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8073c264-8fb4-4316-8b24-5de5ca8238be" containerName="nova-manage" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.374899 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8073c264-8fb4-4316-8b24-5de5ca8238be" containerName="nova-manage" Dec 04 06:32:03 crc kubenswrapper[4685]: E1204 06:32:03.374931 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerName="dnsmasq-dns" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.374938 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerName="dnsmasq-dns" Dec 04 06:32:03 crc kubenswrapper[4685]: E1204 06:32:03.374953 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-metadata" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.374961 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-metadata" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.375266 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c9de0e-75a3-43d1-bbfd-ca70847e1b6d" containerName="dnsmasq-dns" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.375290 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-metadata" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.375327 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b37a166d-763c-4492-bcc8-fef7326f7d82" containerName="nova-cell1-conductor-db-sync" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.375344 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8073c264-8fb4-4316-8b24-5de5ca8238be" containerName="nova-manage" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.375357 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" containerName="nova-metadata-log" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.379822 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.383770 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.384467 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.384867 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.489489 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz6z8\" (UniqueName: \"kubernetes.io/projected/b37a166d-763c-4492-bcc8-fef7326f7d82-kube-api-access-dz6z8\") pod \"b37a166d-763c-4492-bcc8-fef7326f7d82\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.489883 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-combined-ca-bundle\") pod \"b37a166d-763c-4492-bcc8-fef7326f7d82\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.490171 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-scripts\") pod \"b37a166d-763c-4492-bcc8-fef7326f7d82\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.490379 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-config-data\") pod \"b37a166d-763c-4492-bcc8-fef7326f7d82\" (UID: \"b37a166d-763c-4492-bcc8-fef7326f7d82\") " Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.491039 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfjfd\" (UniqueName: \"kubernetes.io/projected/1478f6c7-c54f-4f65-8eab-028f4b5891e9-kube-api-access-sfjfd\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.491162 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.491346 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1478f6c7-c54f-4f65-8eab-028f4b5891e9-logs\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.491684 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.492021 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-config-data\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.495598 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b37a166d-763c-4492-bcc8-fef7326f7d82-kube-api-access-dz6z8" (OuterVolumeSpecName: "kube-api-access-dz6z8") pod "b37a166d-763c-4492-bcc8-fef7326f7d82" (UID: "b37a166d-763c-4492-bcc8-fef7326f7d82"). InnerVolumeSpecName "kube-api-access-dz6z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.500096 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-scripts" (OuterVolumeSpecName: "scripts") pod "b37a166d-763c-4492-bcc8-fef7326f7d82" (UID: "b37a166d-763c-4492-bcc8-fef7326f7d82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.524003 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b37a166d-763c-4492-bcc8-fef7326f7d82" (UID: "b37a166d-763c-4492-bcc8-fef7326f7d82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.526001 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-config-data" (OuterVolumeSpecName: "config-data") pod "b37a166d-763c-4492-bcc8-fef7326f7d82" (UID: "b37a166d-763c-4492-bcc8-fef7326f7d82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594324 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfjfd\" (UniqueName: \"kubernetes.io/projected/1478f6c7-c54f-4f65-8eab-028f4b5891e9-kube-api-access-sfjfd\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594373 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594454 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1478f6c7-c54f-4f65-8eab-028f4b5891e9-logs\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594539 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594650 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-config-data\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594710 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz6z8\" (UniqueName: \"kubernetes.io/projected/b37a166d-763c-4492-bcc8-fef7326f7d82-kube-api-access-dz6z8\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594725 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594738 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.594749 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b37a166d-763c-4492-bcc8-fef7326f7d82-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.595070 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1478f6c7-c54f-4f65-8eab-028f4b5891e9-logs\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.598775 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-config-data\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.599151 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.600290 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.612163 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfjfd\" (UniqueName: \"kubernetes.io/projected/1478f6c7-c54f-4f65-8eab-028f4b5891e9-kube-api-access-sfjfd\") pod \"nova-metadata-0\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.698115 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.973069 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" containerName="nova-scheduler-scheduler" containerID="cri-o://190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" gracePeriod=30 Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.973741 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.973752 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.974768 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zc8bc" event={"ID":"b37a166d-763c-4492-bcc8-fef7326f7d82","Type":"ContainerDied","Data":"e5525345b481ea096d437ba1a91cf0ae0663191b1e5147c4ce57da2953e9f7be"} Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.974796 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5525345b481ea096d437ba1a91cf0ae0663191b1e5147c4ce57da2953e9f7be" Dec 04 06:32:03 crc kubenswrapper[4685]: I1204 06:32:03.974862 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.024583 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.106865 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmh7f\" (UniqueName: \"kubernetes.io/projected/2be52c9a-2015-4e17-a148-525b72b20ee0-kube-api-access-vmh7f\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.106956 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be52c9a-2015-4e17-a148-525b72b20ee0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.107238 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be52c9a-2015-4e17-a148-525b72b20ee0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.199579 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.208774 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmh7f\" (UniqueName: \"kubernetes.io/projected/2be52c9a-2015-4e17-a148-525b72b20ee0-kube-api-access-vmh7f\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.208919 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be52c9a-2015-4e17-a148-525b72b20ee0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.209081 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be52c9a-2015-4e17-a148-525b72b20ee0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.214931 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be52c9a-2015-4e17-a148-525b72b20ee0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.215764 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be52c9a-2015-4e17-a148-525b72b20ee0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.224601 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmh7f\" (UniqueName: \"kubernetes.io/projected/2be52c9a-2015-4e17-a148-525b72b20ee0-kube-api-access-vmh7f\") pod \"nova-cell1-conductor-0\" (UID: \"2be52c9a-2015-4e17-a148-525b72b20ee0\") " pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.323586 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.838730 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.987481 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2be52c9a-2015-4e17-a148-525b72b20ee0","Type":"ContainerStarted","Data":"fdbfff9bb0c6b44ffc261b6767cee512aa32f9948b883d3266864de1390e2987"} Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.991134 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1478f6c7-c54f-4f65-8eab-028f4b5891e9","Type":"ContainerStarted","Data":"41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50"} Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.991235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1478f6c7-c54f-4f65-8eab-028f4b5891e9","Type":"ContainerStarted","Data":"8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67"} Dec 04 06:32:04 crc kubenswrapper[4685]: I1204 06:32:04.991250 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1478f6c7-c54f-4f65-8eab-028f4b5891e9","Type":"ContainerStarted","Data":"927c99a13f562fa0d4e3717408bbb1ee0aed95edd65f0c941e25f721b9e0415a"} Dec 04 06:32:05 crc kubenswrapper[4685]: I1204 06:32:05.024049 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.024032034 podStartE2EDuration="2.024032034s" podCreationTimestamp="2025-12-04 06:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:05.016773064 +0000 UTC m=+1262.205003839" watchObservedRunningTime="2025-12-04 06:32:05.024032034 +0000 UTC m=+1262.212262809" Dec 04 06:32:05 crc kubenswrapper[4685]: I1204 06:32:05.137776 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c870b34c-9387-4fbe-b26a-a487d5bd022a" path="/var/lib/kubelet/pods/c870b34c-9387-4fbe-b26a-a487d5bd022a/volumes" Dec 04 06:32:06 crc kubenswrapper[4685]: I1204 06:32:06.004886 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2be52c9a-2015-4e17-a148-525b72b20ee0","Type":"ContainerStarted","Data":"2f9d76371e1bf160b3afb23f895b5a011006a6e9c16340768c77768098f426b9"} Dec 04 06:32:06 crc kubenswrapper[4685]: I1204 06:32:06.031017 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.030991362 podStartE2EDuration="3.030991362s" podCreationTimestamp="2025-12-04 06:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:06.026895242 +0000 UTC m=+1263.215126047" watchObservedRunningTime="2025-12-04 06:32:06.030991362 +0000 UTC m=+1263.219222167" Dec 04 06:32:06 crc kubenswrapper[4685]: E1204 06:32:06.142614 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 06:32:06 crc kubenswrapper[4685]: E1204 06:32:06.145854 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 06:32:06 crc kubenswrapper[4685]: E1204 06:32:06.148239 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 06:32:06 crc kubenswrapper[4685]: E1204 06:32:06.148296 4685 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" containerName="nova-scheduler-scheduler" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.013209 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.716136 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.868169 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.901742 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-combined-ca-bundle\") pod \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.901852 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-config-data\") pod \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.902017 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd5xq\" (UniqueName: \"kubernetes.io/projected/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-kube-api-access-gd5xq\") pod \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\" (UID: \"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c\") " Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.943580 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-kube-api-access-gd5xq" (OuterVolumeSpecName: "kube-api-access-gd5xq") pod "37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" (UID: "37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c"). InnerVolumeSpecName "kube-api-access-gd5xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.950762 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-config-data" (OuterVolumeSpecName: "config-data") pod "37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" (UID: "37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.952579 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" (UID: "37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:07 crc kubenswrapper[4685]: I1204 06:32:07.965031 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.004331 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.004360 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd5xq\" (UniqueName: \"kubernetes.io/projected/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-kube-api-access-gd5xq\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.004369 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.024996 4685 generic.go:334] "Generic (PLEG): container finished" podID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerID="800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f" exitCode=0 Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.025083 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6f4e7cf1-5021-40ad-9c87-2f0264b339fa","Type":"ContainerDied","Data":"800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f"} Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.025138 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.025154 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6f4e7cf1-5021-40ad-9c87-2f0264b339fa","Type":"ContainerDied","Data":"9daa586a48b3ce4485843d33d837ffa1bb0b945743667c05b31ae173539b5066"} Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.025176 4685 scope.go:117] "RemoveContainer" containerID="800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.028315 4685 generic.go:334] "Generic (PLEG): container finished" podID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" exitCode=0 Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.029106 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.030449 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c","Type":"ContainerDied","Data":"190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31"} Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.030504 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c","Type":"ContainerDied","Data":"35b6473f84ae1a165725df72bf5c71fef6293e3054752de5303acd8617e11fc5"} Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.055306 4685 scope.go:117] "RemoveContainer" containerID="c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.074358 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.079715 4685 scope.go:117] "RemoveContainer" containerID="800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f" Dec 04 06:32:08 crc kubenswrapper[4685]: E1204 06:32:08.083137 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f\": container with ID starting with 800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f not found: ID does not exist" containerID="800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.083201 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f"} err="failed to get container status \"800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f\": rpc error: code = NotFound desc = could not find container \"800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f\": container with ID starting with 800578bd94771afacfbaefcc2f2f902f6c4dbf5ba7fa450701f19c4da3f6980f not found: ID does not exist" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.083235 4685 scope.go:117] "RemoveContainer" containerID="c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b" Dec 04 06:32:08 crc kubenswrapper[4685]: E1204 06:32:08.088656 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b\": container with ID starting with c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b not found: ID does not exist" containerID="c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.088720 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b"} err="failed to get container status \"c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b\": rpc error: code = NotFound desc = could not find container \"c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b\": container with ID starting with c0471b212af1ae129316d848af84201641deb4a3b417edc0d6030b667efacf8b not found: ID does not exist" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.088753 4685 scope.go:117] "RemoveContainer" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.092131 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.101705 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: E1204 06:32:08.103205 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-api" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.103253 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-api" Dec 04 06:32:08 crc kubenswrapper[4685]: E1204 06:32:08.103282 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-log" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.103290 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-log" Dec 04 06:32:08 crc kubenswrapper[4685]: E1204 06:32:08.103332 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" containerName="nova-scheduler-scheduler" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.103341 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" containerName="nova-scheduler-scheduler" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.103960 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" containerName="nova-scheduler-scheduler" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.104001 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-log" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.104036 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" containerName="nova-api-api" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.106701 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9qmv\" (UniqueName: \"kubernetes.io/projected/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-kube-api-access-k9qmv\") pod \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.106863 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-logs\") pod \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.106990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-combined-ca-bundle\") pod \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.107083 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-config-data\") pod \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\" (UID: \"6f4e7cf1-5021-40ad-9c87-2f0264b339fa\") " Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.109308 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.112169 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-logs" (OuterVolumeSpecName: "logs") pod "6f4e7cf1-5021-40ad-9c87-2f0264b339fa" (UID: "6f4e7cf1-5021-40ad-9c87-2f0264b339fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.113055 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.121473 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-kube-api-access-k9qmv" (OuterVolumeSpecName: "kube-api-access-k9qmv") pod "6f4e7cf1-5021-40ad-9c87-2f0264b339fa" (UID: "6f4e7cf1-5021-40ad-9c87-2f0264b339fa"). InnerVolumeSpecName "kube-api-access-k9qmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.132781 4685 scope.go:117] "RemoveContainer" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" Dec 04 06:32:08 crc kubenswrapper[4685]: E1204 06:32:08.133291 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31\": container with ID starting with 190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31 not found: ID does not exist" containerID="190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.133329 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31"} err="failed to get container status \"190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31\": rpc error: code = NotFound desc = could not find container \"190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31\": container with ID starting with 190cfbdf7c1145174e6a48db151e7fe48088c54100f65f49a57bf177af09aa31 not found: ID does not exist" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.146875 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.150668 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-config-data" (OuterVolumeSpecName: "config-data") pod "6f4e7cf1-5021-40ad-9c87-2f0264b339fa" (UID: "6f4e7cf1-5021-40ad-9c87-2f0264b339fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.160073 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f4e7cf1-5021-40ad-9c87-2f0264b339fa" (UID: "6f4e7cf1-5021-40ad-9c87-2f0264b339fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.209977 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-config-data\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.210210 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ml8z\" (UniqueName: \"kubernetes.io/projected/cca78bb8-6393-4be3-8dae-52efe2da3928-kube-api-access-7ml8z\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.211443 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.211637 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.211674 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.211687 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9qmv\" (UniqueName: \"kubernetes.io/projected/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-kube-api-access-k9qmv\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.211702 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4e7cf1-5021-40ad-9c87-2f0264b339fa-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.313236 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.313300 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-config-data\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.313353 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ml8z\" (UniqueName: \"kubernetes.io/projected/cca78bb8-6393-4be3-8dae-52efe2da3928-kube-api-access-7ml8z\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.319585 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.319839 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-config-data\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.329943 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ml8z\" (UniqueName: \"kubernetes.io/projected/cca78bb8-6393-4be3-8dae-52efe2da3928-kube-api-access-7ml8z\") pod \"nova-scheduler-0\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.440303 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.443401 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.460863 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.478076 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.479540 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.481549 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.499933 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.622648 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-config-data\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.622716 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.622755 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwccr\" (UniqueName: \"kubernetes.io/projected/fd914c92-a67e-44ef-9cfa-335cd416f9cf-kube-api-access-dwccr\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.622849 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd914c92-a67e-44ef-9cfa-335cd416f9cf-logs\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.699056 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.699096 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.724357 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd914c92-a67e-44ef-9cfa-335cd416f9cf-logs\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.724466 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-config-data\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.724501 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.724530 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwccr\" (UniqueName: \"kubernetes.io/projected/fd914c92-a67e-44ef-9cfa-335cd416f9cf-kube-api-access-dwccr\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.724743 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd914c92-a67e-44ef-9cfa-335cd416f9cf-logs\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.729863 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-config-data\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.730042 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.739613 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwccr\" (UniqueName: \"kubernetes.io/projected/fd914c92-a67e-44ef-9cfa-335cd416f9cf-kube-api-access-dwccr\") pod \"nova-api-0\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.882733 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:08 crc kubenswrapper[4685]: I1204 06:32:08.968200 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:09 crc kubenswrapper[4685]: I1204 06:32:09.038658 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cca78bb8-6393-4be3-8dae-52efe2da3928","Type":"ContainerStarted","Data":"aa8bbfe727109843bf7423ff1dbd6ad2ba9380b41ee7a501edf1943bd6cd5645"} Dec 04 06:32:09 crc kubenswrapper[4685]: I1204 06:32:09.140573 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c" path="/var/lib/kubelet/pods/37dbad5f-b064-4e4a-a1a4-e3cddcd4c61c/volumes" Dec 04 06:32:09 crc kubenswrapper[4685]: I1204 06:32:09.141132 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f4e7cf1-5021-40ad-9c87-2f0264b339fa" path="/var/lib/kubelet/pods/6f4e7cf1-5021-40ad-9c87-2f0264b339fa/volumes" Dec 04 06:32:09 crc kubenswrapper[4685]: I1204 06:32:09.401501 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:09 crc kubenswrapper[4685]: W1204 06:32:09.406353 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd914c92_a67e_44ef_9cfa_335cd416f9cf.slice/crio-177348b5368222944fd1dbd89825bc72360ee2f426df0b7a605fff510430febe WatchSource:0}: Error finding container 177348b5368222944fd1dbd89825bc72360ee2f426df0b7a605fff510430febe: Status 404 returned error can't find the container with id 177348b5368222944fd1dbd89825bc72360ee2f426df0b7a605fff510430febe Dec 04 06:32:10 crc kubenswrapper[4685]: I1204 06:32:10.063050 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd914c92-a67e-44ef-9cfa-335cd416f9cf","Type":"ContainerStarted","Data":"41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4"} Dec 04 06:32:10 crc kubenswrapper[4685]: I1204 06:32:10.063092 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd914c92-a67e-44ef-9cfa-335cd416f9cf","Type":"ContainerStarted","Data":"4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba"} Dec 04 06:32:10 crc kubenswrapper[4685]: I1204 06:32:10.063101 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd914c92-a67e-44ef-9cfa-335cd416f9cf","Type":"ContainerStarted","Data":"177348b5368222944fd1dbd89825bc72360ee2f426df0b7a605fff510430febe"} Dec 04 06:32:10 crc kubenswrapper[4685]: I1204 06:32:10.064959 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cca78bb8-6393-4be3-8dae-52efe2da3928","Type":"ContainerStarted","Data":"1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95"} Dec 04 06:32:10 crc kubenswrapper[4685]: I1204 06:32:10.084548 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.084531995 podStartE2EDuration="2.084531995s" podCreationTimestamp="2025-12-04 06:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:10.079775405 +0000 UTC m=+1267.268006180" watchObservedRunningTime="2025-12-04 06:32:10.084531995 +0000 UTC m=+1267.272762770" Dec 04 06:32:10 crc kubenswrapper[4685]: I1204 06:32:10.106827 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.106807288 podStartE2EDuration="2.106807288s" podCreationTimestamp="2025-12-04 06:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:10.102954266 +0000 UTC m=+1267.291185041" watchObservedRunningTime="2025-12-04 06:32:10.106807288 +0000 UTC m=+1267.295038063" Dec 04 06:32:12 crc kubenswrapper[4685]: I1204 06:32:12.147256 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:32:12 crc kubenswrapper[4685]: I1204 06:32:12.149251 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9006f3bf-fc55-4204-8e07-df8cae636de4" containerName="kube-state-metrics" containerID="cri-o://c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a" gracePeriod=30 Dec 04 06:32:12 crc kubenswrapper[4685]: I1204 06:32:12.698137 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 06:32:12 crc kubenswrapper[4685]: I1204 06:32:12.808096 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v254c\" (UniqueName: \"kubernetes.io/projected/9006f3bf-fc55-4204-8e07-df8cae636de4-kube-api-access-v254c\") pod \"9006f3bf-fc55-4204-8e07-df8cae636de4\" (UID: \"9006f3bf-fc55-4204-8e07-df8cae636de4\") " Dec 04 06:32:12 crc kubenswrapper[4685]: I1204 06:32:12.818190 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9006f3bf-fc55-4204-8e07-df8cae636de4-kube-api-access-v254c" (OuterVolumeSpecName: "kube-api-access-v254c") pod "9006f3bf-fc55-4204-8e07-df8cae636de4" (UID: "9006f3bf-fc55-4204-8e07-df8cae636de4"). InnerVolumeSpecName "kube-api-access-v254c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:12 crc kubenswrapper[4685]: I1204 06:32:12.911187 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v254c\" (UniqueName: \"kubernetes.io/projected/9006f3bf-fc55-4204-8e07-df8cae636de4-kube-api-access-v254c\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.107027 4685 generic.go:334] "Generic (PLEG): container finished" podID="9006f3bf-fc55-4204-8e07-df8cae636de4" containerID="c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a" exitCode=2 Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.107080 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9006f3bf-fc55-4204-8e07-df8cae636de4","Type":"ContainerDied","Data":"c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a"} Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.107113 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9006f3bf-fc55-4204-8e07-df8cae636de4","Type":"ContainerDied","Data":"d429a5993dc4cc176b1e7f81a6885eab50663f2eb8b24b9c593038f0a0029313"} Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.107119 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.107132 4685 scope.go:117] "RemoveContainer" containerID="c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.136132 4685 scope.go:117] "RemoveContainer" containerID="c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a" Dec 04 06:32:13 crc kubenswrapper[4685]: E1204 06:32:13.139136 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a\": container with ID starting with c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a not found: ID does not exist" containerID="c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.139179 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a"} err="failed to get container status \"c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a\": rpc error: code = NotFound desc = could not find container \"c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a\": container with ID starting with c34f59a45b404ce0b14d4a47b6e9a1a66cb86323b722bc43c2831da100ec458a not found: ID does not exist" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.189562 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.203547 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.213333 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:32:13 crc kubenswrapper[4685]: E1204 06:32:13.213911 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9006f3bf-fc55-4204-8e07-df8cae636de4" containerName="kube-state-metrics" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.213934 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9006f3bf-fc55-4204-8e07-df8cae636de4" containerName="kube-state-metrics" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.214156 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9006f3bf-fc55-4204-8e07-df8cae636de4" containerName="kube-state-metrics" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.215016 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.217314 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.217344 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.219148 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.219192 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.219225 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.219346 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtg6m\" (UniqueName: \"kubernetes.io/projected/a504c958-f275-4b2f-84b7-566b0273f398-kube-api-access-gtg6m\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.224518 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.321450 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtg6m\" (UniqueName: \"kubernetes.io/projected/a504c958-f275-4b2f-84b7-566b0273f398-kube-api-access-gtg6m\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.321531 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.321562 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.321584 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.333461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.333473 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.338663 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a504c958-f275-4b2f-84b7-566b0273f398-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.346448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtg6m\" (UniqueName: \"kubernetes.io/projected/a504c958-f275-4b2f-84b7-566b0273f398-kube-api-access-gtg6m\") pod \"kube-state-metrics-0\" (UID: \"a504c958-f275-4b2f-84b7-566b0273f398\") " pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.440923 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.540239 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.701528 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.703197 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.901989 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.902566 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-central-agent" containerID="cri-o://b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f" gracePeriod=30 Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.903011 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="proxy-httpd" containerID="cri-o://3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793" gracePeriod=30 Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.903046 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-notification-agent" containerID="cri-o://bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439" gracePeriod=30 Dec 04 06:32:13 crc kubenswrapper[4685]: I1204 06:32:13.903123 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="sg-core" containerID="cri-o://5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6" gracePeriod=30 Dec 04 06:32:13 crc kubenswrapper[4685]: E1204 06:32:13.994262 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d8dc337_7239_4721_a3f1_0f97e187865f.slice/crio-5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d8dc337_7239_4721_a3f1_0f97e187865f.slice/crio-conmon-5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.036826 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.128881 4685 generic.go:334] "Generic (PLEG): container finished" podID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerID="5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6" exitCode=2 Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.128963 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerDied","Data":"5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6"} Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.130709 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a504c958-f275-4b2f-84b7-566b0273f398","Type":"ContainerStarted","Data":"c71c20c45ffed2fc8e009414eb39a78f06f0fdcba1ba9fd22f968e86fb346017"} Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.383772 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.723757 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 06:32:14 crc kubenswrapper[4685]: I1204 06:32:14.724373 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.138429 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9006f3bf-fc55-4204-8e07-df8cae636de4" path="/var/lib/kubelet/pods/9006f3bf-fc55-4204-8e07-df8cae636de4/volumes" Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.161253 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a504c958-f275-4b2f-84b7-566b0273f398","Type":"ContainerStarted","Data":"066e2680b763390a616e6bf8db9fd64d78a1342188e6ac660036867827dcd1de"} Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.161561 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.171150 4685 generic.go:334] "Generic (PLEG): container finished" podID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerID="3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793" exitCode=0 Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.171200 4685 generic.go:334] "Generic (PLEG): container finished" podID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerID="b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f" exitCode=0 Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.171235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerDied","Data":"3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793"} Dec 04 06:32:15 crc kubenswrapper[4685]: I1204 06:32:15.171272 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerDied","Data":"b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f"} Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.186511 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.201216 4685 generic.go:334] "Generic (PLEG): container finished" podID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerID="bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439" exitCode=0 Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.201306 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerDied","Data":"bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439"} Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.201370 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d8dc337-7239-4721-a3f1-0f97e187865f","Type":"ContainerDied","Data":"b018329e41742ba013e398e8e5fca7ee14cdbca8b5fdb1220e601811dc428507"} Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.201389 4685 scope.go:117] "RemoveContainer" containerID="3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.201445 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.231353 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.859859531 podStartE2EDuration="4.231333418s" podCreationTimestamp="2025-12-04 06:32:13 +0000 UTC" firstStartedPulling="2025-12-04 06:32:14.040480108 +0000 UTC m=+1271.228710883" lastFinishedPulling="2025-12-04 06:32:14.411953985 +0000 UTC m=+1271.600184770" observedRunningTime="2025-12-04 06:32:15.187551299 +0000 UTC m=+1272.375782104" watchObservedRunningTime="2025-12-04 06:32:17.231333418 +0000 UTC m=+1274.419564203" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.250666 4685 scope.go:117] "RemoveContainer" containerID="5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.256669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-log-httpd\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.256787 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g29zt\" (UniqueName: \"kubernetes.io/projected/0d8dc337-7239-4721-a3f1-0f97e187865f-kube-api-access-g29zt\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.256882 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-config-data\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.256918 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-sg-core-conf-yaml\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.257009 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-run-httpd\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.257097 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-combined-ca-bundle\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.257190 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-scripts\") pod \"0d8dc337-7239-4721-a3f1-0f97e187865f\" (UID: \"0d8dc337-7239-4721-a3f1-0f97e187865f\") " Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.257829 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.258571 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.261027 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.265663 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-scripts" (OuterVolumeSpecName: "scripts") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.266052 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d8dc337-7239-4721-a3f1-0f97e187865f-kube-api-access-g29zt" (OuterVolumeSpecName: "kube-api-access-g29zt") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "kube-api-access-g29zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.291859 4685 scope.go:117] "RemoveContainer" containerID="bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.314274 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.357790 4685 scope.go:117] "RemoveContainer" containerID="b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.360044 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.360081 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.360090 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d8dc337-7239-4721-a3f1-0f97e187865f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.360099 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g29zt\" (UniqueName: \"kubernetes.io/projected/0d8dc337-7239-4721-a3f1-0f97e187865f-kube-api-access-g29zt\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.379456 4685 scope.go:117] "RemoveContainer" containerID="3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.380965 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793\": container with ID starting with 3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793 not found: ID does not exist" containerID="3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.381023 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793"} err="failed to get container status \"3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793\": rpc error: code = NotFound desc = could not find container \"3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793\": container with ID starting with 3c4f0e8916eb9be37ad365463f0889733f20b5470a2967cb835327dbfdd4f793 not found: ID does not exist" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.381055 4685 scope.go:117] "RemoveContainer" containerID="5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.381525 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6\": container with ID starting with 5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6 not found: ID does not exist" containerID="5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.381569 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6"} err="failed to get container status \"5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6\": rpc error: code = NotFound desc = could not find container \"5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6\": container with ID starting with 5861c25245f6713fad5ffeb4921d06e6e0d7a8ee56b96c3041f18168cd0ff6d6 not found: ID does not exist" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.381598 4685 scope.go:117] "RemoveContainer" containerID="bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.381978 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439\": container with ID starting with bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439 not found: ID does not exist" containerID="bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.382016 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439"} err="failed to get container status \"bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439\": rpc error: code = NotFound desc = could not find container \"bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439\": container with ID starting with bac56f12b7b3dbd2e83d24a917b07ca402da13b6b6c7d2b169dd887bd327b439 not found: ID does not exist" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.382040 4685 scope.go:117] "RemoveContainer" containerID="b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.382304 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f\": container with ID starting with b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f not found: ID does not exist" containerID="b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.382335 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f"} err="failed to get container status \"b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f\": rpc error: code = NotFound desc = could not find container \"b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f\": container with ID starting with b8fc9807d93bcc849628597071f58992a3032075e814b44dad3e2fba7f76610f not found: ID does not exist" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.385388 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.391194 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-config-data" (OuterVolumeSpecName: "config-data") pod "0d8dc337-7239-4721-a3f1-0f97e187865f" (UID: "0d8dc337-7239-4721-a3f1-0f97e187865f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.461633 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.461981 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d8dc337-7239-4721-a3f1-0f97e187865f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.547275 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.562497 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.572425 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.573199 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="sg-core" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.573302 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="sg-core" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.573424 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-notification-agent" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.573502 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-notification-agent" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.573595 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-central-agent" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.573668 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-central-agent" Dec 04 06:32:17 crc kubenswrapper[4685]: E1204 06:32:17.573775 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="proxy-httpd" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.573848 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="proxy-httpd" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.574167 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-notification-agent" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.574269 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="proxy-httpd" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.574349 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="sg-core" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.574452 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" containerName="ceilometer-central-agent" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.578944 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.581613 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.581926 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.582920 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.584260 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.666866 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.666929 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-log-httpd\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.666951 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-scripts\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.666971 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.667011 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-run-httpd\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.667034 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-config-data\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.667056 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5jff\" (UniqueName: \"kubernetes.io/projected/1337bdb5-eda1-4d97-a903-6bca4348c8dc-kube-api-access-x5jff\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.667102 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768294 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768359 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768432 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-log-httpd\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768455 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-scripts\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768518 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-run-httpd\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768544 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-config-data\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.768567 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5jff\" (UniqueName: \"kubernetes.io/projected/1337bdb5-eda1-4d97-a903-6bca4348c8dc-kube-api-access-x5jff\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.769441 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-log-httpd\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.771780 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-run-httpd\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.774808 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-config-data\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.775277 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-scripts\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.775531 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.775919 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.778038 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.790208 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5jff\" (UniqueName: \"kubernetes.io/projected/1337bdb5-eda1-4d97-a903-6bca4348c8dc-kube-api-access-x5jff\") pod \"ceilometer-0\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " pod="openstack/ceilometer-0" Dec 04 06:32:17 crc kubenswrapper[4685]: I1204 06:32:17.906149 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:18 crc kubenswrapper[4685]: I1204 06:32:18.365845 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:18 crc kubenswrapper[4685]: I1204 06:32:18.440719 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 06:32:18 crc kubenswrapper[4685]: I1204 06:32:18.477740 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 06:32:18 crc kubenswrapper[4685]: I1204 06:32:18.883091 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 06:32:18 crc kubenswrapper[4685]: I1204 06:32:18.883157 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 06:32:19 crc kubenswrapper[4685]: I1204 06:32:19.139562 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d8dc337-7239-4721-a3f1-0f97e187865f" path="/var/lib/kubelet/pods/0d8dc337-7239-4721-a3f1-0f97e187865f/volumes" Dec 04 06:32:19 crc kubenswrapper[4685]: I1204 06:32:19.224055 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerStarted","Data":"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1"} Dec 04 06:32:19 crc kubenswrapper[4685]: I1204 06:32:19.224107 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerStarted","Data":"b0bf61d99fec9904fd7cf416039e25765e50fcad10ea886143dd717593980bc0"} Dec 04 06:32:19 crc kubenswrapper[4685]: I1204 06:32:19.272732 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 06:32:19 crc kubenswrapper[4685]: I1204 06:32:19.967618 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 06:32:19 crc kubenswrapper[4685]: I1204 06:32:19.967632 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 06:32:20 crc kubenswrapper[4685]: I1204 06:32:20.235446 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerStarted","Data":"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be"} Dec 04 06:32:21 crc kubenswrapper[4685]: I1204 06:32:21.246773 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerStarted","Data":"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015"} Dec 04 06:32:22 crc kubenswrapper[4685]: I1204 06:32:22.260779 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerStarted","Data":"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d"} Dec 04 06:32:22 crc kubenswrapper[4685]: I1204 06:32:22.261373 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:32:22 crc kubenswrapper[4685]: I1204 06:32:22.294286 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.108324392 podStartE2EDuration="5.294260127s" podCreationTimestamp="2025-12-04 06:32:17 +0000 UTC" firstStartedPulling="2025-12-04 06:32:18.369690295 +0000 UTC m=+1275.557921100" lastFinishedPulling="2025-12-04 06:32:21.55562606 +0000 UTC m=+1278.743856835" observedRunningTime="2025-12-04 06:32:22.279279414 +0000 UTC m=+1279.467510189" watchObservedRunningTime="2025-12-04 06:32:22.294260127 +0000 UTC m=+1279.482490932" Dec 04 06:32:23 crc kubenswrapper[4685]: I1204 06:32:23.566820 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 06:32:23 crc kubenswrapper[4685]: I1204 06:32:23.705430 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 06:32:23 crc kubenswrapper[4685]: I1204 06:32:23.705497 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 06:32:23 crc kubenswrapper[4685]: I1204 06:32:23.711589 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 06:32:23 crc kubenswrapper[4685]: I1204 06:32:23.716061 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.243170 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.299807 4685 generic.go:334] "Generic (PLEG): container finished" podID="70d66855-2309-4bd4-9929-794bd475dd3f" containerID="1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03" exitCode=137 Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.299852 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.299854 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d66855-2309-4bd4-9929-794bd475dd3f","Type":"ContainerDied","Data":"1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03"} Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.299963 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d66855-2309-4bd4-9929-794bd475dd3f","Type":"ContainerDied","Data":"2b0426d0873a97db989a91ec526ece6811981fb645dd6b70c6a80764d790c0fd"} Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.299987 4685 scope.go:117] "RemoveContainer" containerID="1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.324618 4685 scope.go:117] "RemoveContainer" containerID="1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03" Dec 04 06:32:26 crc kubenswrapper[4685]: E1204 06:32:26.325497 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03\": container with ID starting with 1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03 not found: ID does not exist" containerID="1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.325537 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03"} err="failed to get container status \"1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03\": rpc error: code = NotFound desc = could not find container \"1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03\": container with ID starting with 1b8fd2fdace212837757f9b55d50e477f71ba6792c94b7cc59f0287036f18b03 not found: ID does not exist" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.336735 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mbsw\" (UniqueName: \"kubernetes.io/projected/70d66855-2309-4bd4-9929-794bd475dd3f-kube-api-access-2mbsw\") pod \"70d66855-2309-4bd4-9929-794bd475dd3f\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.336818 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-combined-ca-bundle\") pod \"70d66855-2309-4bd4-9929-794bd475dd3f\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.336901 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-config-data\") pod \"70d66855-2309-4bd4-9929-794bd475dd3f\" (UID: \"70d66855-2309-4bd4-9929-794bd475dd3f\") " Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.344386 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d66855-2309-4bd4-9929-794bd475dd3f-kube-api-access-2mbsw" (OuterVolumeSpecName: "kube-api-access-2mbsw") pod "70d66855-2309-4bd4-9929-794bd475dd3f" (UID: "70d66855-2309-4bd4-9929-794bd475dd3f"). InnerVolumeSpecName "kube-api-access-2mbsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.372008 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70d66855-2309-4bd4-9929-794bd475dd3f" (UID: "70d66855-2309-4bd4-9929-794bd475dd3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.389177 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-config-data" (OuterVolumeSpecName: "config-data") pod "70d66855-2309-4bd4-9929-794bd475dd3f" (UID: "70d66855-2309-4bd4-9929-794bd475dd3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.440238 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.440310 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mbsw\" (UniqueName: \"kubernetes.io/projected/70d66855-2309-4bd4-9929-794bd475dd3f-kube-api-access-2mbsw\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.440348 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d66855-2309-4bd4-9929-794bd475dd3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.646690 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.656650 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.667971 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:32:26 crc kubenswrapper[4685]: E1204 06:32:26.668467 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d66855-2309-4bd4-9929-794bd475dd3f" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.668488 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d66855-2309-4bd4-9929-794bd475dd3f" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.668840 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d66855-2309-4bd4-9929-794bd475dd3f" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.669550 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.676633 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.676672 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.676707 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.702520 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.745954 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.745997 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l7g7\" (UniqueName: \"kubernetes.io/projected/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-kube-api-access-4l7g7\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.746024 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.746073 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.746105 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.847672 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.847935 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.848092 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.848161 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l7g7\" (UniqueName: \"kubernetes.io/projected/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-kube-api-access-4l7g7\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.848240 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.852325 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.853981 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.854919 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.870175 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.884379 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l7g7\" (UniqueName: \"kubernetes.io/projected/659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f-kube-api-access-4l7g7\") pod \"nova-cell1-novncproxy-0\" (UID: \"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:26 crc kubenswrapper[4685]: I1204 06:32:26.990839 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:27 crc kubenswrapper[4685]: I1204 06:32:27.183088 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d66855-2309-4bd4-9929-794bd475dd3f" path="/var/lib/kubelet/pods/70d66855-2309-4bd4-9929-794bd475dd3f/volumes" Dec 04 06:32:27 crc kubenswrapper[4685]: I1204 06:32:27.524888 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 06:32:27 crc kubenswrapper[4685]: W1204 06:32:27.529642 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod659a5ac2_20c2_4e89_bbb4_c17b6dc90d1f.slice/crio-5f47bd0be087ce0508abdc54d4facbf784dae570a6cdd1f3400181bb95ea0228 WatchSource:0}: Error finding container 5f47bd0be087ce0508abdc54d4facbf784dae570a6cdd1f3400181bb95ea0228: Status 404 returned error can't find the container with id 5f47bd0be087ce0508abdc54d4facbf784dae570a6cdd1f3400181bb95ea0228 Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.322555 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f","Type":"ContainerStarted","Data":"b1bee7eb3c427661b294536a7b0ea97390b556a177ffc5593445ed5ed3a5ab3e"} Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.323139 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f","Type":"ContainerStarted","Data":"5f47bd0be087ce0508abdc54d4facbf784dae570a6cdd1f3400181bb95ea0228"} Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.362679 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.362653956 podStartE2EDuration="2.362653956s" podCreationTimestamp="2025-12-04 06:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:28.35231002 +0000 UTC m=+1285.540540825" watchObservedRunningTime="2025-12-04 06:32:28.362653956 +0000 UTC m=+1285.550884751" Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.888518 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.888642 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.889122 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.889200 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.893339 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 06:32:28 crc kubenswrapper[4685]: I1204 06:32:28.893954 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.147754 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jgt69"] Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.149198 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.167712 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jgt69"] Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.307173 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.307259 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thg6d\" (UniqueName: \"kubernetes.io/projected/12b0951a-f810-4641-a8ba-5d4c528516f4-kube-api-access-thg6d\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.307643 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-config\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.307830 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.308032 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.308104 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.410209 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thg6d\" (UniqueName: \"kubernetes.io/projected/12b0951a-f810-4641-a8ba-5d4c528516f4-kube-api-access-thg6d\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.410257 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-config\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.410650 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.411069 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.411218 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.411287 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.411602 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-config\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.411741 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.412021 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.412309 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.412376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.433136 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thg6d\" (UniqueName: \"kubernetes.io/projected/12b0951a-f810-4641-a8ba-5d4c528516f4-kube-api-access-thg6d\") pod \"dnsmasq-dns-cd5cbd7b9-jgt69\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:29 crc kubenswrapper[4685]: I1204 06:32:29.480778 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:30 crc kubenswrapper[4685]: I1204 06:32:30.002944 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jgt69"] Dec 04 06:32:30 crc kubenswrapper[4685]: W1204 06:32:30.017632 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12b0951a_f810_4641_a8ba_5d4c528516f4.slice/crio-cf838c9a395fa68a8065e963c88f73c0153083bed67094b1b67113912d3dd41f WatchSource:0}: Error finding container cf838c9a395fa68a8065e963c88f73c0153083bed67094b1b67113912d3dd41f: Status 404 returned error can't find the container with id cf838c9a395fa68a8065e963c88f73c0153083bed67094b1b67113912d3dd41f Dec 04 06:32:30 crc kubenswrapper[4685]: I1204 06:32:30.339928 4685 generic.go:334] "Generic (PLEG): container finished" podID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerID="219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5" exitCode=0 Dec 04 06:32:30 crc kubenswrapper[4685]: I1204 06:32:30.340009 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" event={"ID":"12b0951a-f810-4641-a8ba-5d4c528516f4","Type":"ContainerDied","Data":"219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5"} Dec 04 06:32:30 crc kubenswrapper[4685]: I1204 06:32:30.340342 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" event={"ID":"12b0951a-f810-4641-a8ba-5d4c528516f4","Type":"ContainerStarted","Data":"cf838c9a395fa68a8065e963c88f73c0153083bed67094b1b67113912d3dd41f"} Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.016480 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.016839 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-central-agent" containerID="cri-o://a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1" gracePeriod=30 Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.016917 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="sg-core" containerID="cri-o://c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015" gracePeriod=30 Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.016918 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="proxy-httpd" containerID="cri-o://f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d" gracePeriod=30 Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.016954 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-notification-agent" containerID="cri-o://990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be" gracePeriod=30 Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.029335 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": EOF" Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.356557 4685 generic.go:334] "Generic (PLEG): container finished" podID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerID="f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d" exitCode=0 Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.356912 4685 generic.go:334] "Generic (PLEG): container finished" podID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerID="c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015" exitCode=2 Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.356628 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerDied","Data":"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d"} Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.356989 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerDied","Data":"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015"} Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.359890 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" event={"ID":"12b0951a-f810-4641-a8ba-5d4c528516f4","Type":"ContainerStarted","Data":"24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585"} Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.360160 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.391976 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" podStartSLOduration=2.391948676 podStartE2EDuration="2.391948676s" podCreationTimestamp="2025-12-04 06:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:31.381458225 +0000 UTC m=+1288.569689050" watchObservedRunningTime="2025-12-04 06:32:31.391948676 +0000 UTC m=+1288.580179461" Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.990872 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:31 crc kubenswrapper[4685]: I1204 06:32:31.991459 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.067567 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-combined-ca-bundle\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.067677 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-sg-core-conf-yaml\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.067925 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-scripts\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.068025 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-log-httpd\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.068080 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-config-data\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.068842 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-ceilometer-tls-certs\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.068479 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.068956 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-run-httpd\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.069017 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5jff\" (UniqueName: \"kubernetes.io/projected/1337bdb5-eda1-4d97-a903-6bca4348c8dc-kube-api-access-x5jff\") pod \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\" (UID: \"1337bdb5-eda1-4d97-a903-6bca4348c8dc\") " Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.071860 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.075026 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1337bdb5-eda1-4d97-a903-6bca4348c8dc-kube-api-access-x5jff" (OuterVolumeSpecName: "kube-api-access-x5jff") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "kube-api-access-x5jff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.078988 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-scripts" (OuterVolumeSpecName: "scripts") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.079764 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.118540 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.147938 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.174459 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.174492 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.174501 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.174511 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1337bdb5-eda1-4d97-a903-6bca4348c8dc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.174519 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5jff\" (UniqueName: \"kubernetes.io/projected/1337bdb5-eda1-4d97-a903-6bca4348c8dc-kube-api-access-x5jff\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.178650 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.228765 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-config-data" (OuterVolumeSpecName: "config-data") pod "1337bdb5-eda1-4d97-a903-6bca4348c8dc" (UID: "1337bdb5-eda1-4d97-a903-6bca4348c8dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.252262 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.252597 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-log" containerID="cri-o://4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba" gracePeriod=30 Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.252788 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-api" containerID="cri-o://41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4" gracePeriod=30 Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.276766 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.276803 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1337bdb5-eda1-4d97-a903-6bca4348c8dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.371618 4685 generic.go:334] "Generic (PLEG): container finished" podID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerID="990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be" exitCode=0 Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.371646 4685 generic.go:334] "Generic (PLEG): container finished" podID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerID="a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1" exitCode=0 Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.372620 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.372598 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerDied","Data":"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be"} Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.372982 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerDied","Data":"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1"} Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.373109 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1337bdb5-eda1-4d97-a903-6bca4348c8dc","Type":"ContainerDied","Data":"b0bf61d99fec9904fd7cf416039e25765e50fcad10ea886143dd717593980bc0"} Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.373182 4685 scope.go:117] "RemoveContainer" containerID="f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.392761 4685 scope.go:117] "RemoveContainer" containerID="c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.420527 4685 scope.go:117] "RemoveContainer" containerID="990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.422049 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.451336 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.460152 4685 scope.go:117] "RemoveContainer" containerID="a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.476162 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.476745 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-central-agent" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.476824 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-central-agent" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.476897 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="proxy-httpd" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.476957 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="proxy-httpd" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.477018 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-notification-agent" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.477075 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-notification-agent" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.477138 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="sg-core" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.477192 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="sg-core" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.477445 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-central-agent" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.477519 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="proxy-httpd" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.477572 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="sg-core" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.477634 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" containerName="ceilometer-notification-agent" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.479612 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.487279 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.487467 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.487671 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.488759 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.513702 4685 scope.go:117] "RemoveContainer" containerID="f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.515016 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d\": container with ID starting with f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d not found: ID does not exist" containerID="f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.515065 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d"} err="failed to get container status \"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d\": rpc error: code = NotFound desc = could not find container \"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d\": container with ID starting with f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.515095 4685 scope.go:117] "RemoveContainer" containerID="c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.515569 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015\": container with ID starting with c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015 not found: ID does not exist" containerID="c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.515611 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015"} err="failed to get container status \"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015\": rpc error: code = NotFound desc = could not find container \"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015\": container with ID starting with c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015 not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.515636 4685 scope.go:117] "RemoveContainer" containerID="990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.516362 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be\": container with ID starting with 990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be not found: ID does not exist" containerID="990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.516392 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be"} err="failed to get container status \"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be\": rpc error: code = NotFound desc = could not find container \"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be\": container with ID starting with 990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.516475 4685 scope.go:117] "RemoveContainer" containerID="a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1" Dec 04 06:32:32 crc kubenswrapper[4685]: E1204 06:32:32.516723 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1\": container with ID starting with a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1 not found: ID does not exist" containerID="a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.516747 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1"} err="failed to get container status \"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1\": rpc error: code = NotFound desc = could not find container \"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1\": container with ID starting with a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1 not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.516765 4685 scope.go:117] "RemoveContainer" containerID="f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.516989 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d"} err="failed to get container status \"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d\": rpc error: code = NotFound desc = could not find container \"f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d\": container with ID starting with f0186d219543a13e117c995d0aafa18aaadf26b8e29a9e9d8bbf0f58e1dae81d not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.517016 4685 scope.go:117] "RemoveContainer" containerID="c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.517810 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015"} err="failed to get container status \"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015\": rpc error: code = NotFound desc = could not find container \"c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015\": container with ID starting with c446179e87aafa699140fad19c1fb5abc8b2fca2de9627e540f75cf58da4a015 not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.517957 4685 scope.go:117] "RemoveContainer" containerID="990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.518195 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be"} err="failed to get container status \"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be\": rpc error: code = NotFound desc = could not find container \"990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be\": container with ID starting with 990e2ec861a977ac1a9ee292ae5ca217e321793f9d795ea15fc1f6591ecb21be not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.518213 4685 scope.go:117] "RemoveContainer" containerID="a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.518427 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1"} err="failed to get container status \"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1\": rpc error: code = NotFound desc = could not find container \"a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1\": container with ID starting with a8377eb157a0794677eec350d0bf897f4aa2182fc93e5424f6a4377afdfcf3b1 not found: ID does not exist" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.581616 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-scripts\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.581876 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.581975 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9jnq\" (UniqueName: \"kubernetes.io/projected/ac5bc032-6d38-42e0-acda-50d3bc992808-kube-api-access-g9jnq\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.582074 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-log-httpd\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.582174 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-run-httpd\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.582265 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.582389 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.582503 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-config-data\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.683889 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.683933 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9jnq\" (UniqueName: \"kubernetes.io/projected/ac5bc032-6d38-42e0-acda-50d3bc992808-kube-api-access-g9jnq\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.683973 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-log-httpd\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.684004 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-run-httpd\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.684033 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.684071 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.684096 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-config-data\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.684120 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-scripts\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.685146 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-run-httpd\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.685342 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-log-httpd\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.689380 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.689725 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.691019 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-config-data\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.699071 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.699575 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-scripts\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.702683 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9jnq\" (UniqueName: \"kubernetes.io/projected/ac5bc032-6d38-42e0-acda-50d3bc992808-kube-api-access-g9jnq\") pod \"ceilometer-0\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " pod="openstack/ceilometer-0" Dec 04 06:32:32 crc kubenswrapper[4685]: I1204 06:32:32.812858 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:33 crc kubenswrapper[4685]: I1204 06:32:33.137134 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1337bdb5-eda1-4d97-a903-6bca4348c8dc" path="/var/lib/kubelet/pods/1337bdb5-eda1-4d97-a903-6bca4348c8dc/volumes" Dec 04 06:32:33 crc kubenswrapper[4685]: W1204 06:32:33.269655 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac5bc032_6d38_42e0_acda_50d3bc992808.slice/crio-b7b4677f5dee053168dd052f732ab9be2d41b819e885ea79f8c2e7f1296717a7 WatchSource:0}: Error finding container b7b4677f5dee053168dd052f732ab9be2d41b819e885ea79f8c2e7f1296717a7: Status 404 returned error can't find the container with id b7b4677f5dee053168dd052f732ab9be2d41b819e885ea79f8c2e7f1296717a7 Dec 04 06:32:33 crc kubenswrapper[4685]: I1204 06:32:33.269993 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:33 crc kubenswrapper[4685]: I1204 06:32:33.384297 4685 generic.go:334] "Generic (PLEG): container finished" podID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerID="4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba" exitCode=143 Dec 04 06:32:33 crc kubenswrapper[4685]: I1204 06:32:33.384362 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd914c92-a67e-44ef-9cfa-335cd416f9cf","Type":"ContainerDied","Data":"4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba"} Dec 04 06:32:33 crc kubenswrapper[4685]: I1204 06:32:33.385685 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerStarted","Data":"b7b4677f5dee053168dd052f732ab9be2d41b819e885ea79f8c2e7f1296717a7"} Dec 04 06:32:34 crc kubenswrapper[4685]: I1204 06:32:34.402779 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerStarted","Data":"e6731b93de114a01d80b88a98511d10a0d5f2c30258dbf3aedb9ce4cb1ce326e"} Dec 04 06:32:34 crc kubenswrapper[4685]: I1204 06:32:34.741888 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:35 crc kubenswrapper[4685]: I1204 06:32:35.419554 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerStarted","Data":"7fe7858c43ed371c234c4b50d69d69ef98d1176d96633979361e0a5c2d391c58"} Dec 04 06:32:35 crc kubenswrapper[4685]: I1204 06:32:35.419852 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerStarted","Data":"60dc892cd5073c16b6401f33b166c0d89fe1e0b0fa207450841bd001ff509201"} Dec 04 06:32:35 crc kubenswrapper[4685]: I1204 06:32:35.947859 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.043973 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd914c92-a67e-44ef-9cfa-335cd416f9cf-logs\") pod \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.044120 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-combined-ca-bundle\") pod \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.044490 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd914c92-a67e-44ef-9cfa-335cd416f9cf-logs" (OuterVolumeSpecName: "logs") pod "fd914c92-a67e-44ef-9cfa-335cd416f9cf" (UID: "fd914c92-a67e-44ef-9cfa-335cd416f9cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.044900 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwccr\" (UniqueName: \"kubernetes.io/projected/fd914c92-a67e-44ef-9cfa-335cd416f9cf-kube-api-access-dwccr\") pod \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.044947 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-config-data\") pod \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\" (UID: \"fd914c92-a67e-44ef-9cfa-335cd416f9cf\") " Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.045311 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd914c92-a67e-44ef-9cfa-335cd416f9cf-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.052637 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd914c92-a67e-44ef-9cfa-335cd416f9cf-kube-api-access-dwccr" (OuterVolumeSpecName: "kube-api-access-dwccr") pod "fd914c92-a67e-44ef-9cfa-335cd416f9cf" (UID: "fd914c92-a67e-44ef-9cfa-335cd416f9cf"). InnerVolumeSpecName "kube-api-access-dwccr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.074033 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-config-data" (OuterVolumeSpecName: "config-data") pod "fd914c92-a67e-44ef-9cfa-335cd416f9cf" (UID: "fd914c92-a67e-44ef-9cfa-335cd416f9cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.085471 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd914c92-a67e-44ef-9cfa-335cd416f9cf" (UID: "fd914c92-a67e-44ef-9cfa-335cd416f9cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.120768 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.120824 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.146853 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.146891 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwccr\" (UniqueName: \"kubernetes.io/projected/fd914c92-a67e-44ef-9cfa-335cd416f9cf-kube-api-access-dwccr\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.146903 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd914c92-a67e-44ef-9cfa-335cd416f9cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.431961 4685 generic.go:334] "Generic (PLEG): container finished" podID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerID="41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4" exitCode=0 Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.432264 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd914c92-a67e-44ef-9cfa-335cd416f9cf","Type":"ContainerDied","Data":"41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4"} Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.432299 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd914c92-a67e-44ef-9cfa-335cd416f9cf","Type":"ContainerDied","Data":"177348b5368222944fd1dbd89825bc72360ee2f426df0b7a605fff510430febe"} Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.432325 4685 scope.go:117] "RemoveContainer" containerID="41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.432496 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.467639 4685 scope.go:117] "RemoveContainer" containerID="4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.498337 4685 scope.go:117] "RemoveContainer" containerID="41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4" Dec 04 06:32:36 crc kubenswrapper[4685]: E1204 06:32:36.498968 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4\": container with ID starting with 41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4 not found: ID does not exist" containerID="41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.499022 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4"} err="failed to get container status \"41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4\": rpc error: code = NotFound desc = could not find container \"41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4\": container with ID starting with 41ab5d8cf13847a8b22cf1592381f8f5b91ad1b19ff0569e3a14dedd84e2abc4 not found: ID does not exist" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.499116 4685 scope.go:117] "RemoveContainer" containerID="4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba" Dec 04 06:32:36 crc kubenswrapper[4685]: E1204 06:32:36.508272 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba\": container with ID starting with 4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba not found: ID does not exist" containerID="4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.508328 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba"} err="failed to get container status \"4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba\": rpc error: code = NotFound desc = could not find container \"4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba\": container with ID starting with 4849a61bf6971f385628bda6c6abaee9ab0077f6db88896491545ad8948f81ba not found: ID does not exist" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.518582 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.530678 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.544953 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:36 crc kubenswrapper[4685]: E1204 06:32:36.545368 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-log" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.545384 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-log" Dec 04 06:32:36 crc kubenswrapper[4685]: E1204 06:32:36.545422 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-api" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.545430 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-api" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.545608 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-api" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.545640 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" containerName="nova-api-log" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.546596 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.555318 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.561872 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.562061 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.562316 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.563080 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.563128 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-public-tls-certs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.563148 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6730e78-6c4f-4d25-a9d9-e149ec985deb-logs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.563167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686dl\" (UniqueName: \"kubernetes.io/projected/f6730e78-6c4f-4d25-a9d9-e149ec985deb-kube-api-access-686dl\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.563209 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-config-data\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.563274 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.665461 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-public-tls-certs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.665504 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6730e78-6c4f-4d25-a9d9-e149ec985deb-logs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.665527 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686dl\" (UniqueName: \"kubernetes.io/projected/f6730e78-6c4f-4d25-a9d9-e149ec985deb-kube-api-access-686dl\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.665573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-config-data\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.665638 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.665682 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.666501 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6730e78-6c4f-4d25-a9d9-e149ec985deb-logs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.669681 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-config-data\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.669893 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.669922 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-public-tls-certs\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.670552 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.681971 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686dl\" (UniqueName: \"kubernetes.io/projected/f6730e78-6c4f-4d25-a9d9-e149ec985deb-kube-api-access-686dl\") pod \"nova-api-0\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.881737 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:36 crc kubenswrapper[4685]: I1204 06:32:36.992096 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.144442 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd914c92-a67e-44ef-9cfa-335cd416f9cf" path="/var/lib/kubelet/pods/fd914c92-a67e-44ef-9cfa-335cd416f9cf/volumes" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.156139 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.447535 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerStarted","Data":"aba87ec516a07984f37f70b7f214439b0dfa1cb9e9fa78c80ab713cba699c736"} Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.447779 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-central-agent" containerID="cri-o://e6731b93de114a01d80b88a98511d10a0d5f2c30258dbf3aedb9ce4cb1ce326e" gracePeriod=30 Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.448020 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="proxy-httpd" containerID="cri-o://aba87ec516a07984f37f70b7f214439b0dfa1cb9e9fa78c80ab713cba699c736" gracePeriod=30 Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.448061 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-notification-agent" containerID="cri-o://60dc892cd5073c16b6401f33b166c0d89fe1e0b0fa207450841bd001ff509201" gracePeriod=30 Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.448128 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="sg-core" containerID="cri-o://7fe7858c43ed371c234c4b50d69d69ef98d1176d96633979361e0a5c2d391c58" gracePeriod=30 Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.478390 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.40996539 podStartE2EDuration="5.478371065s" podCreationTimestamp="2025-12-04 06:32:32 +0000 UTC" firstStartedPulling="2025-12-04 06:32:33.272910115 +0000 UTC m=+1290.461140890" lastFinishedPulling="2025-12-04 06:32:36.34131578 +0000 UTC m=+1293.529546565" observedRunningTime="2025-12-04 06:32:37.469103462 +0000 UTC m=+1294.657334237" watchObservedRunningTime="2025-12-04 06:32:37.478371065 +0000 UTC m=+1294.666601840" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.485106 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.498046 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:37 crc kubenswrapper[4685]: W1204 06:32:37.505384 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6730e78_6c4f_4d25_a9d9_e149ec985deb.slice/crio-d6b22f16c8636114a647d5aedcd632a9a81180d2b0eea59428889131cbf031f8 WatchSource:0}: Error finding container d6b22f16c8636114a647d5aedcd632a9a81180d2b0eea59428889131cbf031f8: Status 404 returned error can't find the container with id d6b22f16c8636114a647d5aedcd632a9a81180d2b0eea59428889131cbf031f8 Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.702356 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-jzvtg"] Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.703507 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.706702 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.706761 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.713100 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-jzvtg"] Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.892627 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-config-data\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.892921 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.892946 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhzsf\" (UniqueName: \"kubernetes.io/projected/131efcaa-5126-43cc-afa2-264f6c8bee70-kube-api-access-rhzsf\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.893033 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-scripts\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.994321 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-scripts\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.994512 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-config-data\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.994537 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:37 crc kubenswrapper[4685]: I1204 06:32:37.994565 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhzsf\" (UniqueName: \"kubernetes.io/projected/131efcaa-5126-43cc-afa2-264f6c8bee70-kube-api-access-rhzsf\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.000276 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-config-data\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.000729 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-scripts\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.004572 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.019729 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhzsf\" (UniqueName: \"kubernetes.io/projected/131efcaa-5126-43cc-afa2-264f6c8bee70-kube-api-access-rhzsf\") pod \"nova-cell1-cell-mapping-jzvtg\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.028891 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.468031 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6730e78-6c4f-4d25-a9d9-e149ec985deb","Type":"ContainerStarted","Data":"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29"} Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.468717 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6730e78-6c4f-4d25-a9d9-e149ec985deb","Type":"ContainerStarted","Data":"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a"} Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.468730 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6730e78-6c4f-4d25-a9d9-e149ec985deb","Type":"ContainerStarted","Data":"d6b22f16c8636114a647d5aedcd632a9a81180d2b0eea59428889131cbf031f8"} Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.480086 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerID="aba87ec516a07984f37f70b7f214439b0dfa1cb9e9fa78c80ab713cba699c736" exitCode=0 Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.480126 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerID="7fe7858c43ed371c234c4b50d69d69ef98d1176d96633979361e0a5c2d391c58" exitCode=2 Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.480139 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerID="60dc892cd5073c16b6401f33b166c0d89fe1e0b0fa207450841bd001ff509201" exitCode=0 Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.480680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerDied","Data":"aba87ec516a07984f37f70b7f214439b0dfa1cb9e9fa78c80ab713cba699c736"} Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.480768 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerDied","Data":"7fe7858c43ed371c234c4b50d69d69ef98d1176d96633979361e0a5c2d391c58"} Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.480794 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerDied","Data":"60dc892cd5073c16b6401f33b166c0d89fe1e0b0fa207450841bd001ff509201"} Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.486846 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.48682544 podStartE2EDuration="2.48682544s" podCreationTimestamp="2025-12-04 06:32:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:38.485688164 +0000 UTC m=+1295.673918949" watchObservedRunningTime="2025-12-04 06:32:38.48682544 +0000 UTC m=+1295.675056225" Dec 04 06:32:38 crc kubenswrapper[4685]: W1204 06:32:38.507756 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod131efcaa_5126_43cc_afa2_264f6c8bee70.slice/crio-3814027f32ad18b8c3bb136083fb19be2f6daef29cce59f86d8464a28ca92fb2 WatchSource:0}: Error finding container 3814027f32ad18b8c3bb136083fb19be2f6daef29cce59f86d8464a28ca92fb2: Status 404 returned error can't find the container with id 3814027f32ad18b8c3bb136083fb19be2f6daef29cce59f86d8464a28ca92fb2 Dec 04 06:32:38 crc kubenswrapper[4685]: I1204 06:32:38.510642 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-jzvtg"] Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.482598 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.567675 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerID="e6731b93de114a01d80b88a98511d10a0d5f2c30258dbf3aedb9ce4cb1ce326e" exitCode=0 Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.567879 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerDied","Data":"e6731b93de114a01d80b88a98511d10a0d5f2c30258dbf3aedb9ce4cb1ce326e"} Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.580533 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jzvtg" event={"ID":"131efcaa-5126-43cc-afa2-264f6c8bee70","Type":"ContainerStarted","Data":"a2d5a66fde3064601e12843baf9f2145e855b721a21ce447bf3769ad352fc0af"} Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.580583 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jzvtg" event={"ID":"131efcaa-5126-43cc-afa2-264f6c8bee70","Type":"ContainerStarted","Data":"3814027f32ad18b8c3bb136083fb19be2f6daef29cce59f86d8464a28ca92fb2"} Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.602390 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-psqjs"] Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.602695 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" containerName="dnsmasq-dns" containerID="cri-o://d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c" gracePeriod=10 Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.603921 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-jzvtg" podStartSLOduration=2.603902815 podStartE2EDuration="2.603902815s" podCreationTimestamp="2025-12-04 06:32:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:39.602385026 +0000 UTC m=+1296.790615801" watchObservedRunningTime="2025-12-04 06:32:39.603902815 +0000 UTC m=+1296.792133590" Dec 04 06:32:39 crc kubenswrapper[4685]: I1204 06:32:39.990689 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.080874 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9jnq\" (UniqueName: \"kubernetes.io/projected/ac5bc032-6d38-42e0-acda-50d3bc992808-kube-api-access-g9jnq\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.080930 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-scripts\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.082901 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-config-data\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.083149 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-log-httpd\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.083307 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-combined-ca-bundle\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.083486 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-ceilometer-tls-certs\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.084708 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.092806 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-scripts" (OuterVolumeSpecName: "scripts") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.093202 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac5bc032-6d38-42e0-acda-50d3bc992808-kube-api-access-g9jnq" (OuterVolumeSpecName: "kube-api-access-g9jnq") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "kube-api-access-g9jnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.163352 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.188685 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-sg-core-conf-yaml\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.189120 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-run-httpd\") pod \"ac5bc032-6d38-42e0-acda-50d3bc992808\" (UID: \"ac5bc032-6d38-42e0-acda-50d3bc992808\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.189654 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9jnq\" (UniqueName: \"kubernetes.io/projected/ac5bc032-6d38-42e0-acda-50d3bc992808-kube-api-access-g9jnq\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.189678 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.189688 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.189697 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.189652 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.220630 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.226596 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.288643 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-config-data" (OuterVolumeSpecName: "config-data") pod "ac5bc032-6d38-42e0-acda-50d3bc992808" (UID: "ac5bc032-6d38-42e0-acda-50d3bc992808"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.292290 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac5bc032-6d38-42e0-acda-50d3bc992808-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.292316 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.292325 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.292337 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac5bc032-6d38-42e0-acda-50d3bc992808-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.292466 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.393845 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-nb\") pod \"15c9f915-a417-4034-9d34-2a2b1474a104\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.393911 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-swift-storage-0\") pod \"15c9f915-a417-4034-9d34-2a2b1474a104\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.393970 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-svc\") pod \"15c9f915-a417-4034-9d34-2a2b1474a104\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.394042 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-sb\") pod \"15c9f915-a417-4034-9d34-2a2b1474a104\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.394113 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-config\") pod \"15c9f915-a417-4034-9d34-2a2b1474a104\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.394137 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ltwv\" (UniqueName: \"kubernetes.io/projected/15c9f915-a417-4034-9d34-2a2b1474a104-kube-api-access-6ltwv\") pod \"15c9f915-a417-4034-9d34-2a2b1474a104\" (UID: \"15c9f915-a417-4034-9d34-2a2b1474a104\") " Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.397303 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c9f915-a417-4034-9d34-2a2b1474a104-kube-api-access-6ltwv" (OuterVolumeSpecName: "kube-api-access-6ltwv") pod "15c9f915-a417-4034-9d34-2a2b1474a104" (UID: "15c9f915-a417-4034-9d34-2a2b1474a104"). InnerVolumeSpecName "kube-api-access-6ltwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.438956 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15c9f915-a417-4034-9d34-2a2b1474a104" (UID: "15c9f915-a417-4034-9d34-2a2b1474a104"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.441643 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15c9f915-a417-4034-9d34-2a2b1474a104" (UID: "15c9f915-a417-4034-9d34-2a2b1474a104"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.442070 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "15c9f915-a417-4034-9d34-2a2b1474a104" (UID: "15c9f915-a417-4034-9d34-2a2b1474a104"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.443046 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-config" (OuterVolumeSpecName: "config") pod "15c9f915-a417-4034-9d34-2a2b1474a104" (UID: "15c9f915-a417-4034-9d34-2a2b1474a104"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.447285 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15c9f915-a417-4034-9d34-2a2b1474a104" (UID: "15c9f915-a417-4034-9d34-2a2b1474a104"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.496679 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.496710 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.496722 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.496732 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.496740 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c9f915-a417-4034-9d34-2a2b1474a104-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.496749 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ltwv\" (UniqueName: \"kubernetes.io/projected/15c9f915-a417-4034-9d34-2a2b1474a104-kube-api-access-6ltwv\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.593867 4685 generic.go:334] "Generic (PLEG): container finished" podID="15c9f915-a417-4034-9d34-2a2b1474a104" containerID="d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c" exitCode=0 Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.593924 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.593929 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" event={"ID":"15c9f915-a417-4034-9d34-2a2b1474a104","Type":"ContainerDied","Data":"d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c"} Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.594061 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-psqjs" event={"ID":"15c9f915-a417-4034-9d34-2a2b1474a104","Type":"ContainerDied","Data":"e72f144f62838fef8ccde47e8ef7e5ba0ec372cf7f633a98b3f4e4083ae87d25"} Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.594094 4685 scope.go:117] "RemoveContainer" containerID="d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.599474 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac5bc032-6d38-42e0-acda-50d3bc992808","Type":"ContainerDied","Data":"b7b4677f5dee053168dd052f732ab9be2d41b819e885ea79f8c2e7f1296717a7"} Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.599505 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.631873 4685 scope.go:117] "RemoveContainer" containerID="cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.640089 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-psqjs"] Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.651008 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-psqjs"] Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.660295 4685 scope.go:117] "RemoveContainer" containerID="d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.662683 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c\": container with ID starting with d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c not found: ID does not exist" containerID="d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.662754 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c"} err="failed to get container status \"d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c\": rpc error: code = NotFound desc = could not find container \"d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c\": container with ID starting with d20059bdd0defeee117bcde4bec5facae8f4790d36a8df355b33751bb3dc302c not found: ID does not exist" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.662784 4685 scope.go:117] "RemoveContainer" containerID="cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.663182 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819\": container with ID starting with cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819 not found: ID does not exist" containerID="cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.663334 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819"} err="failed to get container status \"cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819\": rpc error: code = NotFound desc = could not find container \"cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819\": container with ID starting with cd9bd63a09822c3cded28bfb92b69975b395d1295e4cd74608557fdced4d8819 not found: ID does not exist" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.663508 4685 scope.go:117] "RemoveContainer" containerID="aba87ec516a07984f37f70b7f214439b0dfa1cb9e9fa78c80ab713cba699c736" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.673065 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.699389 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.701615 4685 scope.go:117] "RemoveContainer" containerID="7fe7858c43ed371c234c4b50d69d69ef98d1176d96633979361e0a5c2d391c58" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.710446 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.710917 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="proxy-httpd" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.710939 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="proxy-httpd" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.710951 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="sg-core" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.710958 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="sg-core" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.710970 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-central-agent" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.710977 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-central-agent" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.710993 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-notification-agent" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711001 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-notification-agent" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.711022 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" containerName="dnsmasq-dns" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711027 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" containerName="dnsmasq-dns" Dec 04 06:32:40 crc kubenswrapper[4685]: E1204 06:32:40.711043 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" containerName="init" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711049 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" containerName="init" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711218 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="sg-core" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711238 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="proxy-httpd" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711256 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-notification-agent" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711263 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" containerName="ceilometer-central-agent" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.711273 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" containerName="dnsmasq-dns" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.713035 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.717102 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.717300 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.718460 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.719431 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.731846 4685 scope.go:117] "RemoveContainer" containerID="60dc892cd5073c16b6401f33b166c0d89fe1e0b0fa207450841bd001ff509201" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.766715 4685 scope.go:117] "RemoveContainer" containerID="e6731b93de114a01d80b88a98511d10a0d5f2c30258dbf3aedb9ce4cb1ce326e" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801743 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801792 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-scripts\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801830 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-config-data\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801857 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/687a62eb-523e-4b35-a54c-aaff0afc1b35-log-httpd\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801878 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801911 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801973 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6tnn\" (UniqueName: \"kubernetes.io/projected/687a62eb-523e-4b35-a54c-aaff0afc1b35-kube-api-access-t6tnn\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.801997 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/687a62eb-523e-4b35-a54c-aaff0afc1b35-run-httpd\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.903971 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-config-data\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.904041 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/687a62eb-523e-4b35-a54c-aaff0afc1b35-log-httpd\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.904107 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.904769 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/687a62eb-523e-4b35-a54c-aaff0afc1b35-log-httpd\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.904856 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.904923 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6tnn\" (UniqueName: \"kubernetes.io/projected/687a62eb-523e-4b35-a54c-aaff0afc1b35-kube-api-access-t6tnn\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.904957 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/687a62eb-523e-4b35-a54c-aaff0afc1b35-run-httpd\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.905284 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/687a62eb-523e-4b35-a54c-aaff0afc1b35-run-httpd\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.905422 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.905537 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-scripts\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.908824 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.909156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.909326 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-scripts\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.911060 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-config-data\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.915806 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687a62eb-523e-4b35-a54c-aaff0afc1b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:40 crc kubenswrapper[4685]: I1204 06:32:40.924965 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6tnn\" (UniqueName: \"kubernetes.io/projected/687a62eb-523e-4b35-a54c-aaff0afc1b35-kube-api-access-t6tnn\") pod \"ceilometer-0\" (UID: \"687a62eb-523e-4b35-a54c-aaff0afc1b35\") " pod="openstack/ceilometer-0" Dec 04 06:32:41 crc kubenswrapper[4685]: I1204 06:32:41.044785 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 06:32:41 crc kubenswrapper[4685]: I1204 06:32:41.143344 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c9f915-a417-4034-9d34-2a2b1474a104" path="/var/lib/kubelet/pods/15c9f915-a417-4034-9d34-2a2b1474a104/volumes" Dec 04 06:32:41 crc kubenswrapper[4685]: I1204 06:32:41.143967 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac5bc032-6d38-42e0-acda-50d3bc992808" path="/var/lib/kubelet/pods/ac5bc032-6d38-42e0-acda-50d3bc992808/volumes" Dec 04 06:32:41 crc kubenswrapper[4685]: I1204 06:32:41.544187 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 06:32:41 crc kubenswrapper[4685]: I1204 06:32:41.611338 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"687a62eb-523e-4b35-a54c-aaff0afc1b35","Type":"ContainerStarted","Data":"765ff1bfd49dbb5f06f0180ac6a2716cbe7327eddd3900fbf42f9822d1b98fc4"} Dec 04 06:32:42 crc kubenswrapper[4685]: I1204 06:32:42.631089 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"687a62eb-523e-4b35-a54c-aaff0afc1b35","Type":"ContainerStarted","Data":"378032658af3212553862ae3a1e2028d2bed0fb8b5a9fb884ea0833355486dd8"} Dec 04 06:32:43 crc kubenswrapper[4685]: I1204 06:32:43.649074 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"687a62eb-523e-4b35-a54c-aaff0afc1b35","Type":"ContainerStarted","Data":"94b149f8287b4b0d0e07b063b7e2a162531df6271202730f3343781dcef71bb4"} Dec 04 06:32:43 crc kubenswrapper[4685]: I1204 06:32:43.653092 4685 generic.go:334] "Generic (PLEG): container finished" podID="131efcaa-5126-43cc-afa2-264f6c8bee70" containerID="a2d5a66fde3064601e12843baf9f2145e855b721a21ce447bf3769ad352fc0af" exitCode=0 Dec 04 06:32:43 crc kubenswrapper[4685]: I1204 06:32:43.653730 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jzvtg" event={"ID":"131efcaa-5126-43cc-afa2-264f6c8bee70","Type":"ContainerDied","Data":"a2d5a66fde3064601e12843baf9f2145e855b721a21ce447bf3769ad352fc0af"} Dec 04 06:32:44 crc kubenswrapper[4685]: I1204 06:32:44.664864 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"687a62eb-523e-4b35-a54c-aaff0afc1b35","Type":"ContainerStarted","Data":"7febe56808f79a137da88ca890ff3f1b5a144652d854f6df2a0b11dc8230c883"} Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.112815 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.202851 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-scripts\") pod \"131efcaa-5126-43cc-afa2-264f6c8bee70\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.203002 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-combined-ca-bundle\") pod \"131efcaa-5126-43cc-afa2-264f6c8bee70\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.203029 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhzsf\" (UniqueName: \"kubernetes.io/projected/131efcaa-5126-43cc-afa2-264f6c8bee70-kube-api-access-rhzsf\") pod \"131efcaa-5126-43cc-afa2-264f6c8bee70\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.203075 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-config-data\") pod \"131efcaa-5126-43cc-afa2-264f6c8bee70\" (UID: \"131efcaa-5126-43cc-afa2-264f6c8bee70\") " Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.207986 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-scripts" (OuterVolumeSpecName: "scripts") pod "131efcaa-5126-43cc-afa2-264f6c8bee70" (UID: "131efcaa-5126-43cc-afa2-264f6c8bee70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.208645 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131efcaa-5126-43cc-afa2-264f6c8bee70-kube-api-access-rhzsf" (OuterVolumeSpecName: "kube-api-access-rhzsf") pod "131efcaa-5126-43cc-afa2-264f6c8bee70" (UID: "131efcaa-5126-43cc-afa2-264f6c8bee70"). InnerVolumeSpecName "kube-api-access-rhzsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.230558 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "131efcaa-5126-43cc-afa2-264f6c8bee70" (UID: "131efcaa-5126-43cc-afa2-264f6c8bee70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.237794 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-config-data" (OuterVolumeSpecName: "config-data") pod "131efcaa-5126-43cc-afa2-264f6c8bee70" (UID: "131efcaa-5126-43cc-afa2-264f6c8bee70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.304207 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.304249 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhzsf\" (UniqueName: \"kubernetes.io/projected/131efcaa-5126-43cc-afa2-264f6c8bee70-kube-api-access-rhzsf\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.304266 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.304278 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131efcaa-5126-43cc-afa2-264f6c8bee70-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.675937 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-jzvtg" event={"ID":"131efcaa-5126-43cc-afa2-264f6c8bee70","Type":"ContainerDied","Data":"3814027f32ad18b8c3bb136083fb19be2f6daef29cce59f86d8464a28ca92fb2"} Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.676209 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3814027f32ad18b8c3bb136083fb19be2f6daef29cce59f86d8464a28ca92fb2" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.676272 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-jzvtg" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.689542 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"687a62eb-523e-4b35-a54c-aaff0afc1b35","Type":"ContainerStarted","Data":"e40952cc4baa46726d262be7e470cc9083badb8d6410681ce5cb3b4d5526334b"} Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.690764 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.728231 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.547749516 podStartE2EDuration="5.728209848s" podCreationTimestamp="2025-12-04 06:32:40 +0000 UTC" firstStartedPulling="2025-12-04 06:32:41.548256504 +0000 UTC m=+1298.736487279" lastFinishedPulling="2025-12-04 06:32:44.728716826 +0000 UTC m=+1301.916947611" observedRunningTime="2025-12-04 06:32:45.719202774 +0000 UTC m=+1302.907433549" watchObservedRunningTime="2025-12-04 06:32:45.728209848 +0000 UTC m=+1302.916440623" Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.865199 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.865477 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cca78bb8-6393-4be3-8dae-52efe2da3928" containerName="nova-scheduler-scheduler" containerID="cri-o://1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95" gracePeriod=30 Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.917745 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.918530 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-log" containerID="cri-o://96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a" gracePeriod=30 Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.918582 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-api" containerID="cri-o://07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29" gracePeriod=30 Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.941572 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.941869 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-log" containerID="cri-o://8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67" gracePeriod=30 Dec 04 06:32:45 crc kubenswrapper[4685]: I1204 06:32:45.942029 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-metadata" containerID="cri-o://41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50" gracePeriod=30 Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.525604 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.638088 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-internal-tls-certs\") pod \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.638242 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-combined-ca-bundle\") pod \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.638330 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-686dl\" (UniqueName: \"kubernetes.io/projected/f6730e78-6c4f-4d25-a9d9-e149ec985deb-kube-api-access-686dl\") pod \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.638538 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6730e78-6c4f-4d25-a9d9-e149ec985deb-logs\") pod \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.638597 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-public-tls-certs\") pod \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.638646 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-config-data\") pod \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\" (UID: \"f6730e78-6c4f-4d25-a9d9-e149ec985deb\") " Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.640079 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6730e78-6c4f-4d25-a9d9-e149ec985deb-logs" (OuterVolumeSpecName: "logs") pod "f6730e78-6c4f-4d25-a9d9-e149ec985deb" (UID: "f6730e78-6c4f-4d25-a9d9-e149ec985deb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.641012 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6730e78-6c4f-4d25-a9d9-e149ec985deb-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.647145 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6730e78-6c4f-4d25-a9d9-e149ec985deb-kube-api-access-686dl" (OuterVolumeSpecName: "kube-api-access-686dl") pod "f6730e78-6c4f-4d25-a9d9-e149ec985deb" (UID: "f6730e78-6c4f-4d25-a9d9-e149ec985deb"). InnerVolumeSpecName "kube-api-access-686dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.670870 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6730e78-6c4f-4d25-a9d9-e149ec985deb" (UID: "f6730e78-6c4f-4d25-a9d9-e149ec985deb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.685516 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-config-data" (OuterVolumeSpecName: "config-data") pod "f6730e78-6c4f-4d25-a9d9-e149ec985deb" (UID: "f6730e78-6c4f-4d25-a9d9-e149ec985deb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.700484 4685 generic.go:334] "Generic (PLEG): container finished" podID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerID="07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29" exitCode=0 Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.700525 4685 generic.go:334] "Generic (PLEG): container finished" podID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerID="96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a" exitCode=143 Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.700629 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.700949 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6730e78-6c4f-4d25-a9d9-e149ec985deb","Type":"ContainerDied","Data":"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29"} Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.701003 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6730e78-6c4f-4d25-a9d9-e149ec985deb","Type":"ContainerDied","Data":"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a"} Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.701016 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6730e78-6c4f-4d25-a9d9-e149ec985deb","Type":"ContainerDied","Data":"d6b22f16c8636114a647d5aedcd632a9a81180d2b0eea59428889131cbf031f8"} Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.701045 4685 scope.go:117] "RemoveContainer" containerID="07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.702851 4685 generic.go:334] "Generic (PLEG): container finished" podID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerID="8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67" exitCode=143 Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.703000 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1478f6c7-c54f-4f65-8eab-028f4b5891e9","Type":"ContainerDied","Data":"8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67"} Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.707505 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f6730e78-6c4f-4d25-a9d9-e149ec985deb" (UID: "f6730e78-6c4f-4d25-a9d9-e149ec985deb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.713700 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f6730e78-6c4f-4d25-a9d9-e149ec985deb" (UID: "f6730e78-6c4f-4d25-a9d9-e149ec985deb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.732753 4685 scope.go:117] "RemoveContainer" containerID="96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.743022 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.743086 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.743099 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-686dl\" (UniqueName: \"kubernetes.io/projected/f6730e78-6c4f-4d25-a9d9-e149ec985deb-kube-api-access-686dl\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.743153 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.743179 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6730e78-6c4f-4d25-a9d9-e149ec985deb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.798637 4685 scope.go:117] "RemoveContainer" containerID="07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29" Dec 04 06:32:46 crc kubenswrapper[4685]: E1204 06:32:46.799114 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29\": container with ID starting with 07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29 not found: ID does not exist" containerID="07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.799148 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29"} err="failed to get container status \"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29\": rpc error: code = NotFound desc = could not find container \"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29\": container with ID starting with 07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29 not found: ID does not exist" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.799184 4685 scope.go:117] "RemoveContainer" containerID="96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a" Dec 04 06:32:46 crc kubenswrapper[4685]: E1204 06:32:46.799806 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a\": container with ID starting with 96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a not found: ID does not exist" containerID="96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.799884 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a"} err="failed to get container status \"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a\": rpc error: code = NotFound desc = could not find container \"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a\": container with ID starting with 96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a not found: ID does not exist" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.799900 4685 scope.go:117] "RemoveContainer" containerID="07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.800134 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29"} err="failed to get container status \"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29\": rpc error: code = NotFound desc = could not find container \"07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29\": container with ID starting with 07330ba522b18f22c7312779b2d3fadcc91c7f0d7bc8f46f7ee6ac03f70b1d29 not found: ID does not exist" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.800170 4685 scope.go:117] "RemoveContainer" containerID="96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a" Dec 04 06:32:46 crc kubenswrapper[4685]: I1204 06:32:46.800441 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a"} err="failed to get container status \"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a\": rpc error: code = NotFound desc = could not find container \"96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a\": container with ID starting with 96874da81434d4d9c6ae8910b4cefc75b3042f218a94ed02b2663d6e7134221a not found: ID does not exist" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.047484 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.054345 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.089928 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: E1204 06:32:47.090551 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-log" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.090579 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-log" Dec 04 06:32:47 crc kubenswrapper[4685]: E1204 06:32:47.090617 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-api" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.090629 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-api" Dec 04 06:32:47 crc kubenswrapper[4685]: E1204 06:32:47.090646 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131efcaa-5126-43cc-afa2-264f6c8bee70" containerName="nova-manage" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.090658 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="131efcaa-5126-43cc-afa2-264f6c8bee70" containerName="nova-manage" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.091046 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="131efcaa-5126-43cc-afa2-264f6c8bee70" containerName="nova-manage" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.091099 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-log" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.091120 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" containerName="nova-api-api" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.101590 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.106141 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.106715 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.110566 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.118063 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.156518 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6730e78-6c4f-4d25-a9d9-e149ec985deb" path="/var/lib/kubelet/pods/f6730e78-6c4f-4d25-a9d9-e149ec985deb/volumes" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.162557 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-public-tls-certs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.162594 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.162714 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-config-data\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.162729 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd4268b9-da6d-496f-96e6-acf27d7e82fa-logs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.162762 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.162974 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pwvx\" (UniqueName: \"kubernetes.io/projected/bd4268b9-da6d-496f-96e6-acf27d7e82fa-kube-api-access-7pwvx\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.265131 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pwvx\" (UniqueName: \"kubernetes.io/projected/bd4268b9-da6d-496f-96e6-acf27d7e82fa-kube-api-access-7pwvx\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.265180 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-public-tls-certs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.265199 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.265253 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-config-data\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.265274 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd4268b9-da6d-496f-96e6-acf27d7e82fa-logs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.265299 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.266023 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd4268b9-da6d-496f-96e6-acf27d7e82fa-logs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.270034 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-config-data\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.271487 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-public-tls-certs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.273669 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.275240 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd4268b9-da6d-496f-96e6-acf27d7e82fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.280478 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pwvx\" (UniqueName: \"kubernetes.io/projected/bd4268b9-da6d-496f-96e6-acf27d7e82fa-kube-api-access-7pwvx\") pod \"nova-api-0\" (UID: \"bd4268b9-da6d-496f-96e6-acf27d7e82fa\") " pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.476756 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.502326 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.672490 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-config-data\") pod \"cca78bb8-6393-4be3-8dae-52efe2da3928\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.672844 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ml8z\" (UniqueName: \"kubernetes.io/projected/cca78bb8-6393-4be3-8dae-52efe2da3928-kube-api-access-7ml8z\") pod \"cca78bb8-6393-4be3-8dae-52efe2da3928\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.672961 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-combined-ca-bundle\") pod \"cca78bb8-6393-4be3-8dae-52efe2da3928\" (UID: \"cca78bb8-6393-4be3-8dae-52efe2da3928\") " Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.680157 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca78bb8-6393-4be3-8dae-52efe2da3928-kube-api-access-7ml8z" (OuterVolumeSpecName: "kube-api-access-7ml8z") pod "cca78bb8-6393-4be3-8dae-52efe2da3928" (UID: "cca78bb8-6393-4be3-8dae-52efe2da3928"). InnerVolumeSpecName "kube-api-access-7ml8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.701887 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cca78bb8-6393-4be3-8dae-52efe2da3928" (UID: "cca78bb8-6393-4be3-8dae-52efe2da3928"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.716252 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-config-data" (OuterVolumeSpecName: "config-data") pod "cca78bb8-6393-4be3-8dae-52efe2da3928" (UID: "cca78bb8-6393-4be3-8dae-52efe2da3928"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.718249 4685 generic.go:334] "Generic (PLEG): container finished" podID="cca78bb8-6393-4be3-8dae-52efe2da3928" containerID="1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95" exitCode=0 Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.718306 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cca78bb8-6393-4be3-8dae-52efe2da3928","Type":"ContainerDied","Data":"1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95"} Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.718335 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cca78bb8-6393-4be3-8dae-52efe2da3928","Type":"ContainerDied","Data":"aa8bbfe727109843bf7423ff1dbd6ad2ba9380b41ee7a501edf1943bd6cd5645"} Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.718351 4685 scope.go:117] "RemoveContainer" containerID="1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.718507 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.742202 4685 scope.go:117] "RemoveContainer" containerID="1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95" Dec 04 06:32:47 crc kubenswrapper[4685]: E1204 06:32:47.743010 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95\": container with ID starting with 1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95 not found: ID does not exist" containerID="1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.743042 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95"} err="failed to get container status \"1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95\": rpc error: code = NotFound desc = could not find container \"1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95\": container with ID starting with 1bcadb772b817dd38c35e67cf6da9995ce1853ef47dbc25ce3df30cf8cd41c95 not found: ID does not exist" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.759870 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.768748 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.774344 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.774374 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cca78bb8-6393-4be3-8dae-52efe2da3928-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.774385 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ml8z\" (UniqueName: \"kubernetes.io/projected/cca78bb8-6393-4be3-8dae-52efe2da3928-kube-api-access-7ml8z\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.788085 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: E1204 06:32:47.788813 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca78bb8-6393-4be3-8dae-52efe2da3928" containerName="nova-scheduler-scheduler" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.788830 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca78bb8-6393-4be3-8dae-52efe2da3928" containerName="nova-scheduler-scheduler" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.789033 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="cca78bb8-6393-4be3-8dae-52efe2da3928" containerName="nova-scheduler-scheduler" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.789642 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.793675 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.815724 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.931120 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.991051 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb3a776-b2f9-448c-b44e-c106669c7c19-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.991090 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjbh\" (UniqueName: \"kubernetes.io/projected/8bb3a776-b2f9-448c-b44e-c106669c7c19-kube-api-access-vsjbh\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:47 crc kubenswrapper[4685]: I1204 06:32:47.991123 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb3a776-b2f9-448c-b44e-c106669c7c19-config-data\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.093486 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb3a776-b2f9-448c-b44e-c106669c7c19-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.093548 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjbh\" (UniqueName: \"kubernetes.io/projected/8bb3a776-b2f9-448c-b44e-c106669c7c19-kube-api-access-vsjbh\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.093589 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb3a776-b2f9-448c-b44e-c106669c7c19-config-data\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.103190 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb3a776-b2f9-448c-b44e-c106669c7c19-config-data\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.106095 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb3a776-b2f9-448c-b44e-c106669c7c19-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.114729 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjbh\" (UniqueName: \"kubernetes.io/projected/8bb3a776-b2f9-448c-b44e-c106669c7c19-kube-api-access-vsjbh\") pod \"nova-scheduler-0\" (UID: \"8bb3a776-b2f9-448c-b44e-c106669c7c19\") " pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.116609 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 06:32:48 crc kubenswrapper[4685]: W1204 06:32:48.664082 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bb3a776_b2f9_448c_b44e_c106669c7c19.slice/crio-8cec9e65519a6cb822c25074c1fd741a0c66dc1c65c452691748ac31ccdba66c WatchSource:0}: Error finding container 8cec9e65519a6cb822c25074c1fd741a0c66dc1c65c452691748ac31ccdba66c: Status 404 returned error can't find the container with id 8cec9e65519a6cb822c25074c1fd741a0c66dc1c65c452691748ac31ccdba66c Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.671622 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.735749 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bb3a776-b2f9-448c-b44e-c106669c7c19","Type":"ContainerStarted","Data":"8cec9e65519a6cb822c25074c1fd741a0c66dc1c65c452691748ac31ccdba66c"} Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.738838 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bd4268b9-da6d-496f-96e6-acf27d7e82fa","Type":"ContainerStarted","Data":"a2893b80cadf306396c6ce5ef64d70ad1edf7454a27b971a7f195387a02293dc"} Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.738874 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bd4268b9-da6d-496f-96e6-acf27d7e82fa","Type":"ContainerStarted","Data":"f1cbee67a9032de4a48eaee3126aa5a4c0b98b8e69d9eaff46df3e69aeb953e6"} Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.738886 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bd4268b9-da6d-496f-96e6-acf27d7e82fa","Type":"ContainerStarted","Data":"b089a6a3503cc1300ac5a80974fa6b6e8e0b5d908468555e7a0b3b5a104b002b"} Dec 04 06:32:48 crc kubenswrapper[4685]: I1204 06:32:48.759046 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.759024677 podStartE2EDuration="1.759024677s" podCreationTimestamp="2025-12-04 06:32:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:48.7566128 +0000 UTC m=+1305.944843575" watchObservedRunningTime="2025-12-04 06:32:48.759024677 +0000 UTC m=+1305.947255452" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.078734 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:32910->10.217.0.191:8775: read: connection reset by peer" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.078846 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:32900->10.217.0.191:8775: read: connection reset by peer" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.137996 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cca78bb8-6393-4be3-8dae-52efe2da3928" path="/var/lib/kubelet/pods/cca78bb8-6393-4be3-8dae-52efe2da3928/volumes" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.613942 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.737098 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-nova-metadata-tls-certs\") pod \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.737154 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-config-data\") pod \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.737266 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfjfd\" (UniqueName: \"kubernetes.io/projected/1478f6c7-c54f-4f65-8eab-028f4b5891e9-kube-api-access-sfjfd\") pod \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.737294 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1478f6c7-c54f-4f65-8eab-028f4b5891e9-logs\") pod \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.737369 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-combined-ca-bundle\") pod \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\" (UID: \"1478f6c7-c54f-4f65-8eab-028f4b5891e9\") " Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.737822 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1478f6c7-c54f-4f65-8eab-028f4b5891e9-logs" (OuterVolumeSpecName: "logs") pod "1478f6c7-c54f-4f65-8eab-028f4b5891e9" (UID: "1478f6c7-c54f-4f65-8eab-028f4b5891e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.754570 4685 generic.go:334] "Generic (PLEG): container finished" podID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerID="41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50" exitCode=0 Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.754627 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1478f6c7-c54f-4f65-8eab-028f4b5891e9","Type":"ContainerDied","Data":"41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50"} Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.754653 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1478f6c7-c54f-4f65-8eab-028f4b5891e9","Type":"ContainerDied","Data":"927c99a13f562fa0d4e3717408bbb1ee0aed95edd65f0c941e25f721b9e0415a"} Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.754669 4685 scope.go:117] "RemoveContainer" containerID="41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.754782 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.760948 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bb3a776-b2f9-448c-b44e-c106669c7c19","Type":"ContainerStarted","Data":"e7223a3d5fb9b5eef8fc64276b36282a42ef53cae720ca76cfccf749ed5f39da"} Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.763227 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1478f6c7-c54f-4f65-8eab-028f4b5891e9-kube-api-access-sfjfd" (OuterVolumeSpecName: "kube-api-access-sfjfd") pod "1478f6c7-c54f-4f65-8eab-028f4b5891e9" (UID: "1478f6c7-c54f-4f65-8eab-028f4b5891e9"). InnerVolumeSpecName "kube-api-access-sfjfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.780266 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1478f6c7-c54f-4f65-8eab-028f4b5891e9" (UID: "1478f6c7-c54f-4f65-8eab-028f4b5891e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.788336 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.78831395 podStartE2EDuration="2.78831395s" podCreationTimestamp="2025-12-04 06:32:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:49.775893407 +0000 UTC m=+1306.964124192" watchObservedRunningTime="2025-12-04 06:32:49.78831395 +0000 UTC m=+1306.976544735" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.792197 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-config-data" (OuterVolumeSpecName: "config-data") pod "1478f6c7-c54f-4f65-8eab-028f4b5891e9" (UID: "1478f6c7-c54f-4f65-8eab-028f4b5891e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.812583 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1478f6c7-c54f-4f65-8eab-028f4b5891e9" (UID: "1478f6c7-c54f-4f65-8eab-028f4b5891e9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.839827 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfjfd\" (UniqueName: \"kubernetes.io/projected/1478f6c7-c54f-4f65-8eab-028f4b5891e9-kube-api-access-sfjfd\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.839858 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1478f6c7-c54f-4f65-8eab-028f4b5891e9-logs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.839869 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.839878 4685 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.839889 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1478f6c7-c54f-4f65-8eab-028f4b5891e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.912285 4685 scope.go:117] "RemoveContainer" containerID="8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.939240 4685 scope.go:117] "RemoveContainer" containerID="41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50" Dec 04 06:32:49 crc kubenswrapper[4685]: E1204 06:32:49.939688 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50\": container with ID starting with 41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50 not found: ID does not exist" containerID="41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.939720 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50"} err="failed to get container status \"41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50\": rpc error: code = NotFound desc = could not find container \"41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50\": container with ID starting with 41875bb89a5c35f8c48e2ad95ae57580d0a670f2c517ed5578f031fbcfdc8f50 not found: ID does not exist" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.939739 4685 scope.go:117] "RemoveContainer" containerID="8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67" Dec 04 06:32:49 crc kubenswrapper[4685]: E1204 06:32:49.941658 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67\": container with ID starting with 8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67 not found: ID does not exist" containerID="8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67" Dec 04 06:32:49 crc kubenswrapper[4685]: I1204 06:32:49.941685 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67"} err="failed to get container status \"8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67\": rpc error: code = NotFound desc = could not find container \"8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67\": container with ID starting with 8ea2fe85b250cac8ee2457b32342cf0a798c2027ee4b6e61bded713ab68f3d67 not found: ID does not exist" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.092616 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.103352 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.122390 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:50 crc kubenswrapper[4685]: E1204 06:32:50.122907 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-log" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.122930 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-log" Dec 04 06:32:50 crc kubenswrapper[4685]: E1204 06:32:50.122972 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-metadata" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.122982 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-metadata" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.123208 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-log" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.123242 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" containerName="nova-metadata-metadata" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.125614 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.130322 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.130632 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.147757 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.248452 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wspgd\" (UniqueName: \"kubernetes.io/projected/104b994d-e75e-4567-9877-f316b73024c4-kube-api-access-wspgd\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.248726 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/104b994d-e75e-4567-9877-f316b73024c4-logs\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.248811 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.248906 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-config-data\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.249088 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.351163 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-config-data\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.352023 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.352205 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wspgd\" (UniqueName: \"kubernetes.io/projected/104b994d-e75e-4567-9877-f316b73024c4-kube-api-access-wspgd\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.352647 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/104b994d-e75e-4567-9877-f316b73024c4-logs\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.353004 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.352964 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/104b994d-e75e-4567-9877-f316b73024c4-logs\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.355396 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-config-data\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.355561 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.357488 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b994d-e75e-4567-9877-f316b73024c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.370657 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wspgd\" (UniqueName: \"kubernetes.io/projected/104b994d-e75e-4567-9877-f316b73024c4-kube-api-access-wspgd\") pod \"nova-metadata-0\" (UID: \"104b994d-e75e-4567-9877-f316b73024c4\") " pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.451213 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 06:32:50 crc kubenswrapper[4685]: I1204 06:32:50.975451 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 06:32:51 crc kubenswrapper[4685]: I1204 06:32:51.151666 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1478f6c7-c54f-4f65-8eab-028f4b5891e9" path="/var/lib/kubelet/pods/1478f6c7-c54f-4f65-8eab-028f4b5891e9/volumes" Dec 04 06:32:51 crc kubenswrapper[4685]: I1204 06:32:51.787199 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"104b994d-e75e-4567-9877-f316b73024c4","Type":"ContainerStarted","Data":"bc758bd2ab1994b807ee69189539f1f83730aa30b7badda0d988390cdca1d7d0"} Dec 04 06:32:51 crc kubenswrapper[4685]: I1204 06:32:51.787680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"104b994d-e75e-4567-9877-f316b73024c4","Type":"ContainerStarted","Data":"f55843e73ea8d9bb444bba1a63e37bec2fc3c361b4cfeba870fc66313a5b3e6e"} Dec 04 06:32:51 crc kubenswrapper[4685]: I1204 06:32:51.787700 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"104b994d-e75e-4567-9877-f316b73024c4","Type":"ContainerStarted","Data":"0adf9e60114b609dadd4cc6fdf8176c42b2829b638178e26b196bd7962e32a6c"} Dec 04 06:32:51 crc kubenswrapper[4685]: I1204 06:32:51.828098 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.828071731 podStartE2EDuration="1.828071731s" podCreationTimestamp="2025-12-04 06:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:32:51.815601048 +0000 UTC m=+1309.003831833" watchObservedRunningTime="2025-12-04 06:32:51.828071731 +0000 UTC m=+1309.016302526" Dec 04 06:32:53 crc kubenswrapper[4685]: I1204 06:32:53.116808 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 06:32:55 crc kubenswrapper[4685]: I1204 06:32:55.452177 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 06:32:55 crc kubenswrapper[4685]: I1204 06:32:55.452676 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 06:32:57 crc kubenswrapper[4685]: I1204 06:32:57.477933 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 06:32:57 crc kubenswrapper[4685]: I1204 06:32:57.478015 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 06:32:58 crc kubenswrapper[4685]: I1204 06:32:58.117580 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 06:32:58 crc kubenswrapper[4685]: I1204 06:32:58.164811 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 06:32:58 crc kubenswrapper[4685]: I1204 06:32:58.496668 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bd4268b9-da6d-496f-96e6-acf27d7e82fa" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 06:32:58 crc kubenswrapper[4685]: I1204 06:32:58.496733 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bd4268b9-da6d-496f-96e6-acf27d7e82fa" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 06:32:58 crc kubenswrapper[4685]: I1204 06:32:58.896919 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 06:33:00 crc kubenswrapper[4685]: I1204 06:33:00.451900 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 06:33:00 crc kubenswrapper[4685]: I1204 06:33:00.453566 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 06:33:01 crc kubenswrapper[4685]: I1204 06:33:01.472723 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="104b994d-e75e-4567-9877-f316b73024c4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 06:33:01 crc kubenswrapper[4685]: I1204 06:33:01.473167 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="104b994d-e75e-4567-9877-f316b73024c4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 06:33:06 crc kubenswrapper[4685]: I1204 06:33:06.121856 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:33:06 crc kubenswrapper[4685]: I1204 06:33:06.122849 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:33:07 crc kubenswrapper[4685]: I1204 06:33:07.489835 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 06:33:07 crc kubenswrapper[4685]: I1204 06:33:07.490390 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 06:33:07 crc kubenswrapper[4685]: I1204 06:33:07.491335 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 06:33:07 crc kubenswrapper[4685]: I1204 06:33:07.491564 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 06:33:07 crc kubenswrapper[4685]: I1204 06:33:07.499685 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 06:33:07 crc kubenswrapper[4685]: I1204 06:33:07.503297 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 06:33:10 crc kubenswrapper[4685]: I1204 06:33:10.461346 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 06:33:10 crc kubenswrapper[4685]: I1204 06:33:10.462263 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 06:33:10 crc kubenswrapper[4685]: I1204 06:33:10.470258 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 06:33:10 crc kubenswrapper[4685]: I1204 06:33:10.999809 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 06:33:11 crc kubenswrapper[4685]: I1204 06:33:11.058317 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 06:33:21 crc kubenswrapper[4685]: I1204 06:33:21.464094 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:33:22 crc kubenswrapper[4685]: I1204 06:33:22.530972 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:33:25 crc kubenswrapper[4685]: I1204 06:33:25.959534 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="rabbitmq" containerID="cri-o://ff8a6270eb509f97464e48636bef08861603eafd47f756d6a55a49213db69fb7" gracePeriod=604796 Dec 04 06:33:26 crc kubenswrapper[4685]: I1204 06:33:26.946384 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="rabbitmq" containerID="cri-o://3245c6f54f8a85b6c7f7f8c68af3d97378f0a05f4933a7250df031df4c1593e9" gracePeriod=604796 Dec 04 06:33:27 crc kubenswrapper[4685]: I1204 06:33:27.876166 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 04 06:33:28 crc kubenswrapper[4685]: I1204 06:33:28.209851 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.254338 4685 generic.go:334] "Generic (PLEG): container finished" podID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerID="ff8a6270eb509f97464e48636bef08861603eafd47f756d6a55a49213db69fb7" exitCode=0 Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.254435 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"777e7995-c510-4d30-bcf1-d479e7cfa432","Type":"ContainerDied","Data":"ff8a6270eb509f97464e48636bef08861603eafd47f756d6a55a49213db69fb7"} Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.518583 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.580724 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-tls\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.580782 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-config-data\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.580823 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ngrc\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-kube-api-access-4ngrc\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.580863 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-plugins\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.580916 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.580994 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-erlang-cookie\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.581012 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-server-conf\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.581043 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/777e7995-c510-4d30-bcf1-d479e7cfa432-pod-info\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.581166 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-confd\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.581199 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-plugins-conf\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.581280 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/777e7995-c510-4d30-bcf1-d479e7cfa432-erlang-cookie-secret\") pod \"777e7995-c510-4d30-bcf1-d479e7cfa432\" (UID: \"777e7995-c510-4d30-bcf1-d479e7cfa432\") " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.582118 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.582487 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.582979 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.589016 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-kube-api-access-4ngrc" (OuterVolumeSpecName: "kube-api-access-4ngrc") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "kube-api-access-4ngrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.592759 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777e7995-c510-4d30-bcf1-d479e7cfa432-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.603654 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.603691 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.606186 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/777e7995-c510-4d30-bcf1-d479e7cfa432-pod-info" (OuterVolumeSpecName: "pod-info") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.644113 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-config-data" (OuterVolumeSpecName: "config-data") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.679825 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-server-conf" (OuterVolumeSpecName: "server-conf") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683703 4685 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683729 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683739 4685 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/777e7995-c510-4d30-bcf1-d479e7cfa432-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683747 4685 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683755 4685 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/777e7995-c510-4d30-bcf1-d479e7cfa432-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683762 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683772 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/777e7995-c510-4d30-bcf1-d479e7cfa432-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683781 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ngrc\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-kube-api-access-4ngrc\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683791 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.683810 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.714632 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.743488 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "777e7995-c510-4d30-bcf1-d479e7cfa432" (UID: "777e7995-c510-4d30-bcf1-d479e7cfa432"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.789597 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:32 crc kubenswrapper[4685]: I1204 06:33:32.789627 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/777e7995-c510-4d30-bcf1-d479e7cfa432-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.265098 4685 generic.go:334] "Generic (PLEG): container finished" podID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerID="3245c6f54f8a85b6c7f7f8c68af3d97378f0a05f4933a7250df031df4c1593e9" exitCode=0 Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.265177 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ff4ac6f-e715-439b-8b4a-2b2715198dab","Type":"ContainerDied","Data":"3245c6f54f8a85b6c7f7f8c68af3d97378f0a05f4933a7250df031df4c1593e9"} Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.269182 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"777e7995-c510-4d30-bcf1-d479e7cfa432","Type":"ContainerDied","Data":"ceecd5d116fa402caf1aeb215c6ca24d49f210c1fe001fadb07df03c65aec1de"} Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.269241 4685 scope.go:117] "RemoveContainer" containerID="ff8a6270eb509f97464e48636bef08861603eafd47f756d6a55a49213db69fb7" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.269264 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.297439 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.311041 4685 scope.go:117] "RemoveContainer" containerID="d20cedfa26345e0d3ffb056841c394db2b49e8dab26dc2b4f19d7034f3ab834f" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.321918 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.338457 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:33:33 crc kubenswrapper[4685]: E1204 06:33:33.338896 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="setup-container" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.338916 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="setup-container" Dec 04 06:33:33 crc kubenswrapper[4685]: E1204 06:33:33.338933 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="rabbitmq" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.338948 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="rabbitmq" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.339142 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" containerName="rabbitmq" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.342053 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.345889 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.350679 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.350924 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.351069 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.351229 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9qfd7" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.351363 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.351526 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.354723 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.414528 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421288 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421354 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421462 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421511 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421531 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-487z7\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-kube-api-access-487z7\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421607 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b87c3420-10be-48df-89b4-c388608be613-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421642 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-config-data\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421696 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421793 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.421854 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b87c3420-10be-48df-89b4-c388608be613-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.523954 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524010 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524044 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524094 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524133 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524158 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-487z7\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-kube-api-access-487z7\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524204 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b87c3420-10be-48df-89b4-c388608be613-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524234 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-config-data\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524267 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524362 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b87c3420-10be-48df-89b4-c388608be613-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.524547 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.525514 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.526189 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.526471 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.526533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.526975 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b87c3420-10be-48df-89b4-c388608be613-config-data\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.530200 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b87c3420-10be-48df-89b4-c388608be613-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.530255 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b87c3420-10be-48df-89b4-c388608be613-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.532903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.543361 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.545533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-487z7\" (UniqueName: \"kubernetes.io/projected/b87c3420-10be-48df-89b4-c388608be613-kube-api-access-487z7\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.574301 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b87c3420-10be-48df-89b4-c388608be613\") " pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.637517 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.728979 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ff4ac6f-e715-439b-8b4a-2b2715198dab-erlang-cookie-secret\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730230 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-plugins-conf\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730301 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730353 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-erlang-cookie\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730436 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-config-data\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730466 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-tls\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730500 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-confd\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730621 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ff4ac6f-e715-439b-8b4a-2b2715198dab-pod-info\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730662 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw244\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-kube-api-access-sw244\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730690 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-plugins\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730794 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-server-conf\") pod \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\" (UID: \"9ff4ac6f-e715-439b-8b4a-2b2715198dab\") " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.730927 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.731385 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.734216 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.734253 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.734723 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.738829 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.745634 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ff4ac6f-e715-439b-8b4a-2b2715198dab-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.748242 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9ff4ac6f-e715-439b-8b4a-2b2715198dab-pod-info" (OuterVolumeSpecName: "pod-info") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.748378 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.753831 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-kube-api-access-sw244" (OuterVolumeSpecName: "kube-api-access-sw244") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "kube-api-access-sw244". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.784991 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-config-data" (OuterVolumeSpecName: "config-data") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.795570 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-server-conf" (OuterVolumeSpecName: "server-conf") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.833773 4685 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834006 4685 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ff4ac6f-e715-439b-8b4a-2b2715198dab-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834020 4685 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834042 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834051 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ff4ac6f-e715-439b-8b4a-2b2715198dab-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834060 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834069 4685 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ff4ac6f-e715-439b-8b4a-2b2715198dab-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834078 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw244\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-kube-api-access-sw244\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.834086 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.864804 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.905110 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9ff4ac6f-e715-439b-8b4a-2b2715198dab" (UID: "9ff4ac6f-e715-439b-8b4a-2b2715198dab"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.935905 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ff4ac6f-e715-439b-8b4a-2b2715198dab-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:33 crc kubenswrapper[4685]: I1204 06:33:33.935929 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.226774 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.294734 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ff4ac6f-e715-439b-8b4a-2b2715198dab","Type":"ContainerDied","Data":"d3162bdd921ea7ebf8c3544ad94fcdb99d442eed9134d0d62ca53d7fc18493fd"} Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.294795 4685 scope.go:117] "RemoveContainer" containerID="3245c6f54f8a85b6c7f7f8c68af3d97378f0a05f4933a7250df031df4c1593e9" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.294882 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.306700 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b87c3420-10be-48df-89b4-c388608be613","Type":"ContainerStarted","Data":"86ad6f1145a6d84e3c271b4a1350c5254e8c1e96810c4703f794842789c03b70"} Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.376078 4685 scope.go:117] "RemoveContainer" containerID="a1c9f7e2cb752c2aae9f83d4ccd671a7be6e02092fac5863bf3f397693e89979" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.419002 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.426599 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.445162 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:33:34 crc kubenswrapper[4685]: E1204 06:33:34.445634 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="rabbitmq" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.445653 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="rabbitmq" Dec 04 06:33:34 crc kubenswrapper[4685]: E1204 06:33:34.445673 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="setup-container" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.445679 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="setup-container" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.445857 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" containerName="rabbitmq" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.446766 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.448986 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.449321 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.449488 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.452226 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.452281 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.452232 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-nlrml" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.452539 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.461118 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.545872 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.545922 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.545973 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546002 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546019 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546049 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546068 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546180 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz9tn\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-kube-api-access-dz9tn\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546306 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546481 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.546517 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648282 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648374 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648437 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648495 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648532 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648571 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz9tn\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-kube-api-access-dz9tn\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648649 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.648678 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649018 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649482 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649521 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649625 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649658 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.649798 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.650614 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.651876 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.668977 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.669310 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.669311 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.669555 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.674172 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz9tn\" (UniqueName: \"kubernetes.io/projected/6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b-kube-api-access-dz9tn\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.705453 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:34 crc kubenswrapper[4685]: I1204 06:33:34.768724 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.145962 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="777e7995-c510-4d30-bcf1-d479e7cfa432" path="/var/lib/kubelet/pods/777e7995-c510-4d30-bcf1-d479e7cfa432/volumes" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.147486 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ff4ac6f-e715-439b-8b4a-2b2715198dab" path="/var/lib/kubelet/pods/9ff4ac6f-e715-439b-8b4a-2b2715198dab/volumes" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.362479 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 06:33:35 crc kubenswrapper[4685]: W1204 06:33:35.373566 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6611a6b2_a9fe_45a2_ad3c_bc30fdd6e08b.slice/crio-776b4b69efeeb4400fc7e331c7d6cf6260d5b5ccd9b3a5df8c33e575eed28bc9 WatchSource:0}: Error finding container 776b4b69efeeb4400fc7e331c7d6cf6260d5b5ccd9b3a5df8c33e575eed28bc9: Status 404 returned error can't find the container with id 776b4b69efeeb4400fc7e331c7d6cf6260d5b5ccd9b3a5df8c33e575eed28bc9 Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.761908 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-bpvtc"] Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.765021 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.766645 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.782783 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-bpvtc"] Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.876957 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.877099 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.877145 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfwdt\" (UniqueName: \"kubernetes.io/projected/2a1a591a-5565-4d69-a063-f65a0542eff5-kube-api-access-sfwdt\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.877167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-svc\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.877217 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.877275 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.877317 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-config\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.978877 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.978957 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.979028 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-config\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.979093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.979182 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.979220 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfwdt\" (UniqueName: \"kubernetes.io/projected/2a1a591a-5565-4d69-a063-f65a0542eff5-kube-api-access-sfwdt\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.979245 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-svc\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.980443 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.980472 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-svc\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.980491 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.980562 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.980664 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-config\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:35 crc kubenswrapper[4685]: I1204 06:33:35.981230 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.068968 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfwdt\" (UniqueName: \"kubernetes.io/projected/2a1a591a-5565-4d69-a063-f65a0542eff5-kube-api-access-sfwdt\") pod \"dnsmasq-dns-d558885bc-bpvtc\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.120727 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.121333 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.121653 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.122495 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6fc1b67a51a8fb7e262f18c32ce16b7ebc33919ea2c7327f8e192aec0ec81fc6"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.122649 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://6fc1b67a51a8fb7e262f18c32ce16b7ebc33919ea2c7327f8e192aec0ec81fc6" gracePeriod=600 Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.337473 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b","Type":"ContainerStarted","Data":"776b4b69efeeb4400fc7e331c7d6cf6260d5b5ccd9b3a5df8c33e575eed28bc9"} Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.339111 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="6fc1b67a51a8fb7e262f18c32ce16b7ebc33919ea2c7327f8e192aec0ec81fc6" exitCode=0 Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.339137 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"6fc1b67a51a8fb7e262f18c32ce16b7ebc33919ea2c7327f8e192aec0ec81fc6"} Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.339171 4685 scope.go:117] "RemoveContainer" containerID="bca40440ad9d82bdf8b013f3ae5e8f16199c54c667e88ee4d71345b125617d55" Dec 04 06:33:36 crc kubenswrapper[4685]: I1204 06:33:36.372239 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:37 crc kubenswrapper[4685]: I1204 06:33:36.853423 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-bpvtc"] Dec 04 06:33:37 crc kubenswrapper[4685]: W1204 06:33:36.858746 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a1a591a_5565_4d69_a063_f65a0542eff5.slice/crio-f75868a40bcf603166c730fe0d0c9356f0dbed33609a68ef5680d2291af5781a WatchSource:0}: Error finding container f75868a40bcf603166c730fe0d0c9356f0dbed33609a68ef5680d2291af5781a: Status 404 returned error can't find the container with id f75868a40bcf603166c730fe0d0c9356f0dbed33609a68ef5680d2291af5781a Dec 04 06:33:37 crc kubenswrapper[4685]: I1204 06:33:37.350035 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47"} Dec 04 06:33:37 crc kubenswrapper[4685]: I1204 06:33:37.351684 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b87c3420-10be-48df-89b4-c388608be613","Type":"ContainerStarted","Data":"eca15741e7360681299463d43a8ff4cc772a2040b774012bd5ee6b68bc6c43cb"} Dec 04 06:33:37 crc kubenswrapper[4685]: I1204 06:33:37.354576 4685 generic.go:334] "Generic (PLEG): container finished" podID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerID="800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc" exitCode=0 Dec 04 06:33:37 crc kubenswrapper[4685]: I1204 06:33:37.354619 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" event={"ID":"2a1a591a-5565-4d69-a063-f65a0542eff5","Type":"ContainerDied","Data":"800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc"} Dec 04 06:33:37 crc kubenswrapper[4685]: I1204 06:33:37.354646 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" event={"ID":"2a1a591a-5565-4d69-a063-f65a0542eff5","Type":"ContainerStarted","Data":"f75868a40bcf603166c730fe0d0c9356f0dbed33609a68ef5680d2291af5781a"} Dec 04 06:33:38 crc kubenswrapper[4685]: I1204 06:33:38.367337 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b","Type":"ContainerStarted","Data":"e645d0143f754a34ef573707d6fff39e646b67819a85808c64d9213f76bde53e"} Dec 04 06:33:38 crc kubenswrapper[4685]: I1204 06:33:38.374613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" event={"ID":"2a1a591a-5565-4d69-a063-f65a0542eff5","Type":"ContainerStarted","Data":"027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af"} Dec 04 06:33:38 crc kubenswrapper[4685]: I1204 06:33:38.423529 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" podStartSLOduration=3.423507304 podStartE2EDuration="3.423507304s" podCreationTimestamp="2025-12-04 06:33:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:33:38.413099005 +0000 UTC m=+1355.601329790" watchObservedRunningTime="2025-12-04 06:33:38.423507304 +0000 UTC m=+1355.611738099" Dec 04 06:33:39 crc kubenswrapper[4685]: I1204 06:33:39.380535 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.375764 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.481819 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jgt69"] Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.482102 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerName="dnsmasq-dns" containerID="cri-o://24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585" gracePeriod=10 Dec 04 06:33:46 crc kubenswrapper[4685]: E1204 06:33:46.590266 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12b0951a_f810_4641_a8ba_5d4c528516f4.slice/crio-conmon-24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.668556 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-vknm9"] Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.686613 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.708863 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-vknm9"] Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.818536 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.818620 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.818790 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.818854 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.818920 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-config\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.819049 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.819154 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trmbs\" (UniqueName: \"kubernetes.io/projected/22be434f-4bd4-4e59-b980-ae992dd4cf7d-kube-api-access-trmbs\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921635 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921683 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921718 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-config\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921749 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921770 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trmbs\" (UniqueName: \"kubernetes.io/projected/22be434f-4bd4-4e59-b980-ae992dd4cf7d-kube-api-access-trmbs\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921853 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.921905 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.922714 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.922807 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.923294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.923401 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.923461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.923551 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be434f-4bd4-4e59-b980-ae992dd4cf7d-config\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:46 crc kubenswrapper[4685]: I1204 06:33:46.941334 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trmbs\" (UniqueName: \"kubernetes.io/projected/22be434f-4bd4-4e59-b980-ae992dd4cf7d-kube-api-access-trmbs\") pod \"dnsmasq-dns-78c64bc9c5-vknm9\" (UID: \"22be434f-4bd4-4e59-b980-ae992dd4cf7d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.014795 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.029578 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.125822 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-config\") pod \"12b0951a-f810-4641-a8ba-5d4c528516f4\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.125872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-sb\") pod \"12b0951a-f810-4641-a8ba-5d4c528516f4\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.125894 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-svc\") pod \"12b0951a-f810-4641-a8ba-5d4c528516f4\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.125970 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-nb\") pod \"12b0951a-f810-4641-a8ba-5d4c528516f4\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.126023 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thg6d\" (UniqueName: \"kubernetes.io/projected/12b0951a-f810-4641-a8ba-5d4c528516f4-kube-api-access-thg6d\") pod \"12b0951a-f810-4641-a8ba-5d4c528516f4\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.126165 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-swift-storage-0\") pod \"12b0951a-f810-4641-a8ba-5d4c528516f4\" (UID: \"12b0951a-f810-4641-a8ba-5d4c528516f4\") " Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.131227 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12b0951a-f810-4641-a8ba-5d4c528516f4-kube-api-access-thg6d" (OuterVolumeSpecName: "kube-api-access-thg6d") pod "12b0951a-f810-4641-a8ba-5d4c528516f4" (UID: "12b0951a-f810-4641-a8ba-5d4c528516f4"). InnerVolumeSpecName "kube-api-access-thg6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.173352 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "12b0951a-f810-4641-a8ba-5d4c528516f4" (UID: "12b0951a-f810-4641-a8ba-5d4c528516f4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.173805 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "12b0951a-f810-4641-a8ba-5d4c528516f4" (UID: "12b0951a-f810-4641-a8ba-5d4c528516f4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.190514 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "12b0951a-f810-4641-a8ba-5d4c528516f4" (UID: "12b0951a-f810-4641-a8ba-5d4c528516f4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.206553 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "12b0951a-f810-4641-a8ba-5d4c528516f4" (UID: "12b0951a-f810-4641-a8ba-5d4c528516f4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.208522 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-config" (OuterVolumeSpecName: "config") pod "12b0951a-f810-4641-a8ba-5d4c528516f4" (UID: "12b0951a-f810-4641-a8ba-5d4c528516f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.237286 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.237312 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.237323 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.237331 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.237340 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thg6d\" (UniqueName: \"kubernetes.io/projected/12b0951a-f810-4641-a8ba-5d4c528516f4-kube-api-access-thg6d\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.237350 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/12b0951a-f810-4641-a8ba-5d4c528516f4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.478574 4685 generic.go:334] "Generic (PLEG): container finished" podID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerID="24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585" exitCode=0 Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.478621 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" event={"ID":"12b0951a-f810-4641-a8ba-5d4c528516f4","Type":"ContainerDied","Data":"24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585"} Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.478648 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" event={"ID":"12b0951a-f810-4641-a8ba-5d4c528516f4","Type":"ContainerDied","Data":"cf838c9a395fa68a8065e963c88f73c0153083bed67094b1b67113912d3dd41f"} Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.478663 4685 scope.go:117] "RemoveContainer" containerID="24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.478762 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jgt69" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.533526 4685 scope.go:117] "RemoveContainer" containerID="219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.533608 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-vknm9"] Dec 04 06:33:47 crc kubenswrapper[4685]: W1204 06:33:47.539160 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22be434f_4bd4_4e59_b980_ae992dd4cf7d.slice/crio-bee085884073ef0ab11be824a2cb43725bfbe53ad75beb07039adf07afce9cef WatchSource:0}: Error finding container bee085884073ef0ab11be824a2cb43725bfbe53ad75beb07039adf07afce9cef: Status 404 returned error can't find the container with id bee085884073ef0ab11be824a2cb43725bfbe53ad75beb07039adf07afce9cef Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.542619 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jgt69"] Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.550773 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jgt69"] Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.555134 4685 scope.go:117] "RemoveContainer" containerID="24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585" Dec 04 06:33:47 crc kubenswrapper[4685]: E1204 06:33:47.555612 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585\": container with ID starting with 24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585 not found: ID does not exist" containerID="24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.555791 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585"} err="failed to get container status \"24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585\": rpc error: code = NotFound desc = could not find container \"24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585\": container with ID starting with 24e7ed090af97852240c77ff21b23ce14f2ad078a91805a44522efc645963585 not found: ID does not exist" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.555939 4685 scope.go:117] "RemoveContainer" containerID="219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5" Dec 04 06:33:47 crc kubenswrapper[4685]: E1204 06:33:47.556378 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5\": container with ID starting with 219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5 not found: ID does not exist" containerID="219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5" Dec 04 06:33:47 crc kubenswrapper[4685]: I1204 06:33:47.556421 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5"} err="failed to get container status \"219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5\": rpc error: code = NotFound desc = could not find container \"219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5\": container with ID starting with 219e3347139aec6aab740fa528c84047e8d4b6c9227f9bab41f25a69d50196a5 not found: ID does not exist" Dec 04 06:33:48 crc kubenswrapper[4685]: I1204 06:33:48.501982 4685 generic.go:334] "Generic (PLEG): container finished" podID="22be434f-4bd4-4e59-b980-ae992dd4cf7d" containerID="3c0d04565c1b9a867d652986d319abec61fe2fb76224685af920bb407b6705e1" exitCode=0 Dec 04 06:33:48 crc kubenswrapper[4685]: I1204 06:33:48.502739 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" event={"ID":"22be434f-4bd4-4e59-b980-ae992dd4cf7d","Type":"ContainerDied","Data":"3c0d04565c1b9a867d652986d319abec61fe2fb76224685af920bb407b6705e1"} Dec 04 06:33:48 crc kubenswrapper[4685]: I1204 06:33:48.502832 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" event={"ID":"22be434f-4bd4-4e59-b980-ae992dd4cf7d","Type":"ContainerStarted","Data":"bee085884073ef0ab11be824a2cb43725bfbe53ad75beb07039adf07afce9cef"} Dec 04 06:33:49 crc kubenswrapper[4685]: I1204 06:33:49.147055 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" path="/var/lib/kubelet/pods/12b0951a-f810-4641-a8ba-5d4c528516f4/volumes" Dec 04 06:33:49 crc kubenswrapper[4685]: I1204 06:33:49.520233 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" event={"ID":"22be434f-4bd4-4e59-b980-ae992dd4cf7d","Type":"ContainerStarted","Data":"7b3f0160328fa9e887dc695ee173ae9b97a2d196d0968ad534b62ccba8ffe516"} Dec 04 06:33:49 crc kubenswrapper[4685]: I1204 06:33:49.520514 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:49 crc kubenswrapper[4685]: I1204 06:33:49.551338 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" podStartSLOduration=3.551311712 podStartE2EDuration="3.551311712s" podCreationTimestamp="2025-12-04 06:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:33:49.545878204 +0000 UTC m=+1366.734109019" watchObservedRunningTime="2025-12-04 06:33:49.551311712 +0000 UTC m=+1366.739542527" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.017334 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-vknm9" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.109195 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-bpvtc"] Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.109460 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerName="dnsmasq-dns" containerID="cri-o://027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af" gracePeriod=10 Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.610255 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.624792 4685 generic.go:334] "Generic (PLEG): container finished" podID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerID="027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af" exitCode=0 Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.624849 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.624840 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" event={"ID":"2a1a591a-5565-4d69-a063-f65a0542eff5","Type":"ContainerDied","Data":"027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af"} Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.624919 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-bpvtc" event={"ID":"2a1a591a-5565-4d69-a063-f65a0542eff5","Type":"ContainerDied","Data":"f75868a40bcf603166c730fe0d0c9356f0dbed33609a68ef5680d2291af5781a"} Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.624938 4685 scope.go:117] "RemoveContainer" containerID="027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.660982 4685 scope.go:117] "RemoveContainer" containerID="800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683353 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-nb\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-sb\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683515 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfwdt\" (UniqueName: \"kubernetes.io/projected/2a1a591a-5565-4d69-a063-f65a0542eff5-kube-api-access-sfwdt\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683549 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-swift-storage-0\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683607 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-openstack-edpm-ipam\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683707 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-config\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.683738 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-svc\") pod \"2a1a591a-5565-4d69-a063-f65a0542eff5\" (UID: \"2a1a591a-5565-4d69-a063-f65a0542eff5\") " Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.687963 4685 scope.go:117] "RemoveContainer" containerID="027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af" Dec 04 06:33:57 crc kubenswrapper[4685]: E1204 06:33:57.688896 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af\": container with ID starting with 027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af not found: ID does not exist" containerID="027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.688947 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af"} err="failed to get container status \"027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af\": rpc error: code = NotFound desc = could not find container \"027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af\": container with ID starting with 027cc14dcff705d2f46dcefb99482005b23abb8912ec3c57e13abdf899b1a9af not found: ID does not exist" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.688974 4685 scope.go:117] "RemoveContainer" containerID="800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc" Dec 04 06:33:57 crc kubenswrapper[4685]: E1204 06:33:57.689342 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc\": container with ID starting with 800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc not found: ID does not exist" containerID="800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.689382 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc"} err="failed to get container status \"800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc\": rpc error: code = NotFound desc = could not find container \"800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc\": container with ID starting with 800eca18763d7bb9d6bcd8edda56d046c076b32b8393c679e29e3fcf5ea9f0fc not found: ID does not exist" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.690098 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a1a591a-5565-4d69-a063-f65a0542eff5-kube-api-access-sfwdt" (OuterVolumeSpecName: "kube-api-access-sfwdt") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "kube-api-access-sfwdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.735782 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-config" (OuterVolumeSpecName: "config") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.741394 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.741809 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.742923 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.748103 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.771122 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2a1a591a-5565-4d69-a063-f65a0542eff5" (UID: "2a1a591a-5565-4d69-a063-f65a0542eff5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785469 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785496 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-config\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785505 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785519 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785529 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785537 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfwdt\" (UniqueName: \"kubernetes.io/projected/2a1a591a-5565-4d69-a063-f65a0542eff5-kube-api-access-sfwdt\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.785547 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a1a591a-5565-4d69-a063-f65a0542eff5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.980545 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-bpvtc"] Dec 04 06:33:57 crc kubenswrapper[4685]: I1204 06:33:57.993477 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-bpvtc"] Dec 04 06:33:59 crc kubenswrapper[4685]: I1204 06:33:59.141245 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" path="/var/lib/kubelet/pods/2a1a591a-5565-4d69-a063-f65a0542eff5/volumes" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.455576 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd"] Dec 04 06:34:09 crc kubenswrapper[4685]: E1204 06:34:09.456687 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerName="init" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.456702 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerName="init" Dec 04 06:34:09 crc kubenswrapper[4685]: E1204 06:34:09.456726 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerName="dnsmasq-dns" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.456735 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerName="dnsmasq-dns" Dec 04 06:34:09 crc kubenswrapper[4685]: E1204 06:34:09.456744 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerName="dnsmasq-dns" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.456752 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerName="dnsmasq-dns" Dec 04 06:34:09 crc kubenswrapper[4685]: E1204 06:34:09.456773 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerName="init" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.456780 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerName="init" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.457005 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="12b0951a-f810-4641-a8ba-5d4c528516f4" containerName="dnsmasq-dns" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.457032 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a1a591a-5565-4d69-a063-f65a0542eff5" containerName="dnsmasq-dns" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.457866 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.460340 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.460535 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.461035 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.461755 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.471006 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd"] Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.567945 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrrwk\" (UniqueName: \"kubernetes.io/projected/7d609a36-8df9-42ab-8160-c6f8b49880d9-kube-api-access-rrrwk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.568167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.568240 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.568318 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.671164 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.671451 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.671665 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.671769 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrrwk\" (UniqueName: \"kubernetes.io/projected/7d609a36-8df9-42ab-8160-c6f8b49880d9-kube-api-access-rrrwk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.680565 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.680966 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.681213 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.705171 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrrwk\" (UniqueName: \"kubernetes.io/projected/7d609a36-8df9-42ab-8160-c6f8b49880d9-kube-api-access-rrrwk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.788732 4685 generic.go:334] "Generic (PLEG): container finished" podID="b87c3420-10be-48df-89b4-c388608be613" containerID="eca15741e7360681299463d43a8ff4cc772a2040b774012bd5ee6b68bc6c43cb" exitCode=0 Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.789541 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b87c3420-10be-48df-89b4-c388608be613","Type":"ContainerDied","Data":"eca15741e7360681299463d43a8ff4cc772a2040b774012bd5ee6b68bc6c43cb"} Dec 04 06:34:09 crc kubenswrapper[4685]: I1204 06:34:09.795747 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:10 crc kubenswrapper[4685]: I1204 06:34:10.417399 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd"] Dec 04 06:34:10 crc kubenswrapper[4685]: W1204 06:34:10.423508 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d609a36_8df9_42ab_8160_c6f8b49880d9.slice/crio-b708d79af98332d023901fc7c12fce426c50480fba407b09a436b7f247e15ca4 WatchSource:0}: Error finding container b708d79af98332d023901fc7c12fce426c50480fba407b09a436b7f247e15ca4: Status 404 returned error can't find the container with id b708d79af98332d023901fc7c12fce426c50480fba407b09a436b7f247e15ca4 Dec 04 06:34:10 crc kubenswrapper[4685]: I1204 06:34:10.432402 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:34:10 crc kubenswrapper[4685]: I1204 06:34:10.813269 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b87c3420-10be-48df-89b4-c388608be613","Type":"ContainerStarted","Data":"50f728e6a72aba899689c7230e9d1ee18a0c5779e239873637dff1a893bc0530"} Dec 04 06:34:10 crc kubenswrapper[4685]: I1204 06:34:10.813602 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 06:34:10 crc kubenswrapper[4685]: I1204 06:34:10.823899 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" event={"ID":"7d609a36-8df9-42ab-8160-c6f8b49880d9","Type":"ContainerStarted","Data":"b708d79af98332d023901fc7c12fce426c50480fba407b09a436b7f247e15ca4"} Dec 04 06:34:10 crc kubenswrapper[4685]: I1204 06:34:10.845471 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.845454124 podStartE2EDuration="37.845454124s" podCreationTimestamp="2025-12-04 06:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:34:10.839881272 +0000 UTC m=+1388.028112047" watchObservedRunningTime="2025-12-04 06:34:10.845454124 +0000 UTC m=+1388.033684899" Dec 04 06:34:11 crc kubenswrapper[4685]: I1204 06:34:11.836945 4685 generic.go:334] "Generic (PLEG): container finished" podID="6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b" containerID="e645d0143f754a34ef573707d6fff39e646b67819a85808c64d9213f76bde53e" exitCode=0 Dec 04 06:34:11 crc kubenswrapper[4685]: I1204 06:34:11.837547 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b","Type":"ContainerDied","Data":"e645d0143f754a34ef573707d6fff39e646b67819a85808c64d9213f76bde53e"} Dec 04 06:34:12 crc kubenswrapper[4685]: I1204 06:34:12.848986 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b","Type":"ContainerStarted","Data":"04da90a60e8dee86c34cf48151800df62f58936c06186281fa7f121f66bd4a89"} Dec 04 06:34:12 crc kubenswrapper[4685]: I1204 06:34:12.849260 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:34:12 crc kubenswrapper[4685]: I1204 06:34:12.873878 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.873858445 podStartE2EDuration="38.873858445s" podCreationTimestamp="2025-12-04 06:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:34:12.872575725 +0000 UTC m=+1390.060806510" watchObservedRunningTime="2025-12-04 06:34:12.873858445 +0000 UTC m=+1390.062089210" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.775770 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-875xs"] Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.778658 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.798737 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-875xs"] Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.819983 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-utilities\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.820139 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-catalog-content\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.820165 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5ngl\" (UniqueName: \"kubernetes.io/projected/dc60e06f-e78e-4f53-9e9b-af826e3b8447-kube-api-access-g5ngl\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.922273 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-catalog-content\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.922328 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5ngl\" (UniqueName: \"kubernetes.io/projected/dc60e06f-e78e-4f53-9e9b-af826e3b8447-kube-api-access-g5ngl\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.922423 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-utilities\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.922903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-catalog-content\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.922955 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-utilities\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:15 crc kubenswrapper[4685]: I1204 06:34:15.945224 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5ngl\" (UniqueName: \"kubernetes.io/projected/dc60e06f-e78e-4f53-9e9b-af826e3b8447-kube-api-access-g5ngl\") pod \"redhat-marketplace-875xs\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:16 crc kubenswrapper[4685]: I1204 06:34:16.109924 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:16 crc kubenswrapper[4685]: I1204 06:34:16.583888 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-875xs"] Dec 04 06:34:16 crc kubenswrapper[4685]: I1204 06:34:16.889193 4685 generic.go:334] "Generic (PLEG): container finished" podID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerID="df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073" exitCode=0 Dec 04 06:34:16 crc kubenswrapper[4685]: I1204 06:34:16.889466 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerDied","Data":"df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073"} Dec 04 06:34:16 crc kubenswrapper[4685]: I1204 06:34:16.889496 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerStarted","Data":"5c25d2d958e8faf28f3d09f40b4ffb1ee28bd2a83a6f10f7f50141f8d21e8052"} Dec 04 06:34:17 crc kubenswrapper[4685]: I1204 06:34:17.899903 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerStarted","Data":"1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f"} Dec 04 06:34:18 crc kubenswrapper[4685]: I1204 06:34:18.912281 4685 generic.go:334] "Generic (PLEG): container finished" podID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerID="1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f" exitCode=0 Dec 04 06:34:18 crc kubenswrapper[4685]: I1204 06:34:18.912343 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerDied","Data":"1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f"} Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.153918 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-smx4x"] Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.156686 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.166808 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-smx4x"] Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.241254 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t47dx\" (UniqueName: \"kubernetes.io/projected/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-kube-api-access-t47dx\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.241307 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-catalog-content\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.244507 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-utilities\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.346784 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t47dx\" (UniqueName: \"kubernetes.io/projected/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-kube-api-access-t47dx\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.346872 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-catalog-content\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.346982 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-utilities\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.347395 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-catalog-content\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.347882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-utilities\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.374794 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t47dx\" (UniqueName: \"kubernetes.io/projected/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-kube-api-access-t47dx\") pod \"redhat-operators-smx4x\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:21 crc kubenswrapper[4685]: I1204 06:34:21.489958 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:23 crc kubenswrapper[4685]: I1204 06:34:23.737679 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 06:34:24 crc kubenswrapper[4685]: I1204 06:34:24.773852 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 06:34:25 crc kubenswrapper[4685]: I1204 06:34:25.823331 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-smx4x"] Dec 04 06:34:25 crc kubenswrapper[4685]: I1204 06:34:25.988801 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerStarted","Data":"a9886cd4fc437261d2aa705c9188c6f9d1dc35587cec1dcfde3c5227150096a2"} Dec 04 06:34:26 crc kubenswrapper[4685]: I1204 06:34:26.999632 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerStarted","Data":"770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c"} Dec 04 06:34:27 crc kubenswrapper[4685]: I1204 06:34:27.001262 4685 generic.go:334] "Generic (PLEG): container finished" podID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerID="dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea" exitCode=0 Dec 04 06:34:27 crc kubenswrapper[4685]: I1204 06:34:27.001293 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerDied","Data":"dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea"} Dec 04 06:34:27 crc kubenswrapper[4685]: I1204 06:34:27.004490 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" event={"ID":"7d609a36-8df9-42ab-8160-c6f8b49880d9","Type":"ContainerStarted","Data":"043a5781f7107a89d9072d0acd43c220358cc7e12474dde32f959e4e4af1dd36"} Dec 04 06:34:27 crc kubenswrapper[4685]: I1204 06:34:27.023324 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-875xs" podStartSLOduration=3.100157484 podStartE2EDuration="12.023305836s" podCreationTimestamp="2025-12-04 06:34:15 +0000 UTC" firstStartedPulling="2025-12-04 06:34:16.891088524 +0000 UTC m=+1394.079319309" lastFinishedPulling="2025-12-04 06:34:25.814236886 +0000 UTC m=+1403.002467661" observedRunningTime="2025-12-04 06:34:27.020757627 +0000 UTC m=+1404.208988402" watchObservedRunningTime="2025-12-04 06:34:27.023305836 +0000 UTC m=+1404.211536611" Dec 04 06:34:27 crc kubenswrapper[4685]: I1204 06:34:27.051981 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" podStartSLOduration=2.638129051 podStartE2EDuration="18.051957961s" podCreationTimestamp="2025-12-04 06:34:09 +0000 UTC" firstStartedPulling="2025-12-04 06:34:10.432109995 +0000 UTC m=+1387.620340780" lastFinishedPulling="2025-12-04 06:34:25.845938915 +0000 UTC m=+1403.034169690" observedRunningTime="2025-12-04 06:34:27.041873679 +0000 UTC m=+1404.230104464" watchObservedRunningTime="2025-12-04 06:34:27.051957961 +0000 UTC m=+1404.240188756" Dec 04 06:34:29 crc kubenswrapper[4685]: I1204 06:34:29.029552 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerStarted","Data":"96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b"} Dec 04 06:34:30 crc kubenswrapper[4685]: I1204 06:34:30.046175 4685 generic.go:334] "Generic (PLEG): container finished" podID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerID="96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b" exitCode=0 Dec 04 06:34:30 crc kubenswrapper[4685]: I1204 06:34:30.046218 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerDied","Data":"96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b"} Dec 04 06:34:31 crc kubenswrapper[4685]: I1204 06:34:31.058486 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerStarted","Data":"dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a"} Dec 04 06:34:31 crc kubenswrapper[4685]: I1204 06:34:31.082229 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-smx4x" podStartSLOduration=6.395551905 podStartE2EDuration="10.082205802s" podCreationTimestamp="2025-12-04 06:34:21 +0000 UTC" firstStartedPulling="2025-12-04 06:34:27.003900667 +0000 UTC m=+1404.192131442" lastFinishedPulling="2025-12-04 06:34:30.690554554 +0000 UTC m=+1407.878785339" observedRunningTime="2025-12-04 06:34:31.078192318 +0000 UTC m=+1408.266423133" watchObservedRunningTime="2025-12-04 06:34:31.082205802 +0000 UTC m=+1408.270436567" Dec 04 06:34:31 crc kubenswrapper[4685]: I1204 06:34:31.490963 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:31 crc kubenswrapper[4685]: I1204 06:34:31.491116 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:32 crc kubenswrapper[4685]: I1204 06:34:32.546357 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-smx4x" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="registry-server" probeResult="failure" output=< Dec 04 06:34:32 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Dec 04 06:34:32 crc kubenswrapper[4685]: > Dec 04 06:34:36 crc kubenswrapper[4685]: I1204 06:34:36.110753 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:36 crc kubenswrapper[4685]: I1204 06:34:36.111498 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:36 crc kubenswrapper[4685]: I1204 06:34:36.173717 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:37 crc kubenswrapper[4685]: I1204 06:34:37.122169 4685 generic.go:334] "Generic (PLEG): container finished" podID="7d609a36-8df9-42ab-8160-c6f8b49880d9" containerID="043a5781f7107a89d9072d0acd43c220358cc7e12474dde32f959e4e4af1dd36" exitCode=0 Dec 04 06:34:37 crc kubenswrapper[4685]: I1204 06:34:37.122256 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" event={"ID":"7d609a36-8df9-42ab-8160-c6f8b49880d9","Type":"ContainerDied","Data":"043a5781f7107a89d9072d0acd43c220358cc7e12474dde32f959e4e4af1dd36"} Dec 04 06:34:37 crc kubenswrapper[4685]: I1204 06:34:37.179754 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:37 crc kubenswrapper[4685]: I1204 06:34:37.239077 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-875xs"] Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.593749 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.667394 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-inventory\") pod \"7d609a36-8df9-42ab-8160-c6f8b49880d9\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.667571 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrrwk\" (UniqueName: \"kubernetes.io/projected/7d609a36-8df9-42ab-8160-c6f8b49880d9-kube-api-access-rrrwk\") pod \"7d609a36-8df9-42ab-8160-c6f8b49880d9\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.667777 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-repo-setup-combined-ca-bundle\") pod \"7d609a36-8df9-42ab-8160-c6f8b49880d9\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.667843 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-ssh-key\") pod \"7d609a36-8df9-42ab-8160-c6f8b49880d9\" (UID: \"7d609a36-8df9-42ab-8160-c6f8b49880d9\") " Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.683701 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d609a36-8df9-42ab-8160-c6f8b49880d9-kube-api-access-rrrwk" (OuterVolumeSpecName: "kube-api-access-rrrwk") pod "7d609a36-8df9-42ab-8160-c6f8b49880d9" (UID: "7d609a36-8df9-42ab-8160-c6f8b49880d9"). InnerVolumeSpecName "kube-api-access-rrrwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.687469 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7d609a36-8df9-42ab-8160-c6f8b49880d9" (UID: "7d609a36-8df9-42ab-8160-c6f8b49880d9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.718673 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7d609a36-8df9-42ab-8160-c6f8b49880d9" (UID: "7d609a36-8df9-42ab-8160-c6f8b49880d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.722189 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-inventory" (OuterVolumeSpecName: "inventory") pod "7d609a36-8df9-42ab-8160-c6f8b49880d9" (UID: "7d609a36-8df9-42ab-8160-c6f8b49880d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.770560 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrrwk\" (UniqueName: \"kubernetes.io/projected/7d609a36-8df9-42ab-8160-c6f8b49880d9-kube-api-access-rrrwk\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.770602 4685 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.770616 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:38 crc kubenswrapper[4685]: I1204 06:34:38.770629 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d609a36-8df9-42ab-8160-c6f8b49880d9-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.148059 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.148100 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd" event={"ID":"7d609a36-8df9-42ab-8160-c6f8b49880d9","Type":"ContainerDied","Data":"b708d79af98332d023901fc7c12fce426c50480fba407b09a436b7f247e15ca4"} Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.148139 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b708d79af98332d023901fc7c12fce426c50480fba407b09a436b7f247e15ca4" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.148188 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-875xs" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="registry-server" containerID="cri-o://770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c" gracePeriod=2 Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.250234 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n"] Dec 04 06:34:39 crc kubenswrapper[4685]: E1204 06:34:39.250750 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d609a36-8df9-42ab-8160-c6f8b49880d9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.250773 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d609a36-8df9-42ab-8160-c6f8b49880d9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.251032 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d609a36-8df9-42ab-8160-c6f8b49880d9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.251834 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.254054 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.254437 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.256813 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.257703 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.262827 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n"] Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.281209 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.281291 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpg86\" (UniqueName: \"kubernetes.io/projected/08644b36-c6a1-4c4f-9e4d-89e861f2c197-kube-api-access-fpg86\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.281323 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.383237 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.383553 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpg86\" (UniqueName: \"kubernetes.io/projected/08644b36-c6a1-4c4f-9e4d-89e861f2c197-kube-api-access-fpg86\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.383619 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.387294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.387640 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.398773 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpg86\" (UniqueName: \"kubernetes.io/projected/08644b36-c6a1-4c4f-9e4d-89e861f2c197-kube-api-access-fpg86\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tx98n\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.611953 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.635003 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.688239 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-catalog-content\") pod \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.688408 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5ngl\" (UniqueName: \"kubernetes.io/projected/dc60e06f-e78e-4f53-9e9b-af826e3b8447-kube-api-access-g5ngl\") pod \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.688681 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-utilities\") pod \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\" (UID: \"dc60e06f-e78e-4f53-9e9b-af826e3b8447\") " Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.689870 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-utilities" (OuterVolumeSpecName: "utilities") pod "dc60e06f-e78e-4f53-9e9b-af826e3b8447" (UID: "dc60e06f-e78e-4f53-9e9b-af826e3b8447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.699683 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc60e06f-e78e-4f53-9e9b-af826e3b8447-kube-api-access-g5ngl" (OuterVolumeSpecName: "kube-api-access-g5ngl") pod "dc60e06f-e78e-4f53-9e9b-af826e3b8447" (UID: "dc60e06f-e78e-4f53-9e9b-af826e3b8447"). InnerVolumeSpecName "kube-api-access-g5ngl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.741830 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc60e06f-e78e-4f53-9e9b-af826e3b8447" (UID: "dc60e06f-e78e-4f53-9e9b-af826e3b8447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.790592 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.790860 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc60e06f-e78e-4f53-9e9b-af826e3b8447-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:39 crc kubenswrapper[4685]: I1204 06:34:39.790872 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5ngl\" (UniqueName: \"kubernetes.io/projected/dc60e06f-e78e-4f53-9e9b-af826e3b8447-kube-api-access-g5ngl\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.160255 4685 generic.go:334] "Generic (PLEG): container finished" podID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerID="770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c" exitCode=0 Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.160300 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerDied","Data":"770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c"} Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.160321 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-875xs" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.160335 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-875xs" event={"ID":"dc60e06f-e78e-4f53-9e9b-af826e3b8447","Type":"ContainerDied","Data":"5c25d2d958e8faf28f3d09f40b4ffb1ee28bd2a83a6f10f7f50141f8d21e8052"} Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.160357 4685 scope.go:117] "RemoveContainer" containerID="770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.202968 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-875xs"] Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.206892 4685 scope.go:117] "RemoveContainer" containerID="1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.213768 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-875xs"] Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.231237 4685 scope.go:117] "RemoveContainer" containerID="df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.306954 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n"] Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.313189 4685 scope.go:117] "RemoveContainer" containerID="770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c" Dec 04 06:34:40 crc kubenswrapper[4685]: E1204 06:34:40.313625 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c\": container with ID starting with 770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c not found: ID does not exist" containerID="770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.313669 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c"} err="failed to get container status \"770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c\": rpc error: code = NotFound desc = could not find container \"770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c\": container with ID starting with 770ac86feb97ccb7f4f51cd4d659641e80f8fb2c88c3b0489612873c5b62f93c not found: ID does not exist" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.313695 4685 scope.go:117] "RemoveContainer" containerID="1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f" Dec 04 06:34:40 crc kubenswrapper[4685]: E1204 06:34:40.314228 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f\": container with ID starting with 1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f not found: ID does not exist" containerID="1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.314265 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f"} err="failed to get container status \"1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f\": rpc error: code = NotFound desc = could not find container \"1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f\": container with ID starting with 1e495119ca548a89fc4b02642d8b054206a35f0eb615ff8cf4acf1025f34626f not found: ID does not exist" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.314292 4685 scope.go:117] "RemoveContainer" containerID="df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073" Dec 04 06:34:40 crc kubenswrapper[4685]: E1204 06:34:40.314677 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073\": container with ID starting with df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073 not found: ID does not exist" containerID="df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073" Dec 04 06:34:40 crc kubenswrapper[4685]: I1204 06:34:40.314703 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073"} err="failed to get container status \"df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073\": rpc error: code = NotFound desc = could not find container \"df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073\": container with ID starting with df3b322e44a7f1fa43c4186376edea817f1e7341c2bec00f444da03005f9a073 not found: ID does not exist" Dec 04 06:34:40 crc kubenswrapper[4685]: W1204 06:34:40.319500 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08644b36_c6a1_4c4f_9e4d_89e861f2c197.slice/crio-ced52fa4114255fbded05bc59efa6c0e8ecc67c5e2d38553f836f6658c7ed36a WatchSource:0}: Error finding container ced52fa4114255fbded05bc59efa6c0e8ecc67c5e2d38553f836f6658c7ed36a: Status 404 returned error can't find the container with id ced52fa4114255fbded05bc59efa6c0e8ecc67c5e2d38553f836f6658c7ed36a Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.141492 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" path="/var/lib/kubelet/pods/dc60e06f-e78e-4f53-9e9b-af826e3b8447/volumes" Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.182660 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" event={"ID":"08644b36-c6a1-4c4f-9e4d-89e861f2c197","Type":"ContainerStarted","Data":"46bbaf829778ebc9f2c6331a31685a787bb588c8d0eb6d02354d8154ab4347d9"} Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.182714 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" event={"ID":"08644b36-c6a1-4c4f-9e4d-89e861f2c197","Type":"ContainerStarted","Data":"ced52fa4114255fbded05bc59efa6c0e8ecc67c5e2d38553f836f6658c7ed36a"} Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.206586 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" podStartSLOduration=2.012060552 podStartE2EDuration="2.206567141s" podCreationTimestamp="2025-12-04 06:34:39 +0000 UTC" firstStartedPulling="2025-12-04 06:34:40.324872954 +0000 UTC m=+1417.513103749" lastFinishedPulling="2025-12-04 06:34:40.519379553 +0000 UTC m=+1417.707610338" observedRunningTime="2025-12-04 06:34:41.204218608 +0000 UTC m=+1418.392449393" watchObservedRunningTime="2025-12-04 06:34:41.206567141 +0000 UTC m=+1418.394797916" Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.567638 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.610916 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:41 crc kubenswrapper[4685]: I1204 06:34:41.736690 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-smx4x"] Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.202784 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-smx4x" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="registry-server" containerID="cri-o://dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a" gracePeriod=2 Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.706375 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.773252 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-utilities\") pod \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.773715 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t47dx\" (UniqueName: \"kubernetes.io/projected/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-kube-api-access-t47dx\") pod \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.773798 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-catalog-content\") pod \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\" (UID: \"5fe94ddc-d25d-4020-b08c-4e01c6f0055f\") " Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.774626 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-utilities" (OuterVolumeSpecName: "utilities") pod "5fe94ddc-d25d-4020-b08c-4e01c6f0055f" (UID: "5fe94ddc-d25d-4020-b08c-4e01c6f0055f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.779627 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-kube-api-access-t47dx" (OuterVolumeSpecName: "kube-api-access-t47dx") pod "5fe94ddc-d25d-4020-b08c-4e01c6f0055f" (UID: "5fe94ddc-d25d-4020-b08c-4e01c6f0055f"). InnerVolumeSpecName "kube-api-access-t47dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.876828 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.876866 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t47dx\" (UniqueName: \"kubernetes.io/projected/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-kube-api-access-t47dx\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.926513 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fe94ddc-d25d-4020-b08c-4e01c6f0055f" (UID: "5fe94ddc-d25d-4020-b08c-4e01c6f0055f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:34:43 crc kubenswrapper[4685]: I1204 06:34:43.978245 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe94ddc-d25d-4020-b08c-4e01c6f0055f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.211008 4685 generic.go:334] "Generic (PLEG): container finished" podID="08644b36-c6a1-4c4f-9e4d-89e861f2c197" containerID="46bbaf829778ebc9f2c6331a31685a787bb588c8d0eb6d02354d8154ab4347d9" exitCode=0 Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.211079 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" event={"ID":"08644b36-c6a1-4c4f-9e4d-89e861f2c197","Type":"ContainerDied","Data":"46bbaf829778ebc9f2c6331a31685a787bb588c8d0eb6d02354d8154ab4347d9"} Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.213801 4685 generic.go:334] "Generic (PLEG): container finished" podID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerID="dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a" exitCode=0 Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.213844 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerDied","Data":"dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a"} Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.213899 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smx4x" event={"ID":"5fe94ddc-d25d-4020-b08c-4e01c6f0055f","Type":"ContainerDied","Data":"a9886cd4fc437261d2aa705c9188c6f9d1dc35587cec1dcfde3c5227150096a2"} Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.213917 4685 scope.go:117] "RemoveContainer" containerID="dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.213988 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smx4x" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.243649 4685 scope.go:117] "RemoveContainer" containerID="96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.249584 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-smx4x"] Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.258615 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-smx4x"] Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.285193 4685 scope.go:117] "RemoveContainer" containerID="dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.320551 4685 scope.go:117] "RemoveContainer" containerID="dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a" Dec 04 06:34:44 crc kubenswrapper[4685]: E1204 06:34:44.321041 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a\": container with ID starting with dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a not found: ID does not exist" containerID="dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.321080 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a"} err="failed to get container status \"dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a\": rpc error: code = NotFound desc = could not find container \"dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a\": container with ID starting with dd6290b8226d55760c8e099ee7212bb169b79d67dd5859cdd3d89c3803da610a not found: ID does not exist" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.321100 4685 scope.go:117] "RemoveContainer" containerID="96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b" Dec 04 06:34:44 crc kubenswrapper[4685]: E1204 06:34:44.321448 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b\": container with ID starting with 96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b not found: ID does not exist" containerID="96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.321488 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b"} err="failed to get container status \"96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b\": rpc error: code = NotFound desc = could not find container \"96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b\": container with ID starting with 96fc128e7f32ad72a733266308fa70c875fca0c898bc1aedae5be16029acd71b not found: ID does not exist" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.321517 4685 scope.go:117] "RemoveContainer" containerID="dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea" Dec 04 06:34:44 crc kubenswrapper[4685]: E1204 06:34:44.321866 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea\": container with ID starting with dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea not found: ID does not exist" containerID="dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea" Dec 04 06:34:44 crc kubenswrapper[4685]: I1204 06:34:44.321890 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea"} err="failed to get container status \"dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea\": rpc error: code = NotFound desc = could not find container \"dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea\": container with ID starting with dc9373d44f913c0288936529704d7d9b54ad2582ad39bf024b0d4053557771ea not found: ID does not exist" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.140363 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" path="/var/lib/kubelet/pods/5fe94ddc-d25d-4020-b08c-4e01c6f0055f/volumes" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.700289 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.813968 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-inventory\") pod \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.814078 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpg86\" (UniqueName: \"kubernetes.io/projected/08644b36-c6a1-4c4f-9e4d-89e861f2c197-kube-api-access-fpg86\") pod \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.814202 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-ssh-key\") pod \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\" (UID: \"08644b36-c6a1-4c4f-9e4d-89e861f2c197\") " Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.820978 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08644b36-c6a1-4c4f-9e4d-89e861f2c197-kube-api-access-fpg86" (OuterVolumeSpecName: "kube-api-access-fpg86") pod "08644b36-c6a1-4c4f-9e4d-89e861f2c197" (UID: "08644b36-c6a1-4c4f-9e4d-89e861f2c197"). InnerVolumeSpecName "kube-api-access-fpg86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.859581 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-inventory" (OuterVolumeSpecName: "inventory") pod "08644b36-c6a1-4c4f-9e4d-89e861f2c197" (UID: "08644b36-c6a1-4c4f-9e4d-89e861f2c197"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.868748 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08644b36-c6a1-4c4f-9e4d-89e861f2c197" (UID: "08644b36-c6a1-4c4f-9e4d-89e861f2c197"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.916565 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.916607 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpg86\" (UniqueName: \"kubernetes.io/projected/08644b36-c6a1-4c4f-9e4d-89e861f2c197-kube-api-access-fpg86\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:45 crc kubenswrapper[4685]: I1204 06:34:45.916628 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08644b36-c6a1-4c4f-9e4d-89e861f2c197-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.242196 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" event={"ID":"08644b36-c6a1-4c4f-9e4d-89e861f2c197","Type":"ContainerDied","Data":"ced52fa4114255fbded05bc59efa6c0e8ecc67c5e2d38553f836f6658c7ed36a"} Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.242262 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ced52fa4114255fbded05bc59efa6c0e8ecc67c5e2d38553f836f6658c7ed36a" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.242292 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tx98n" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334382 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2"] Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334839 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="extract-utilities" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334854 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="extract-utilities" Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334885 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="extract-content" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334894 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="extract-content" Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334907 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="registry-server" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334916 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="registry-server" Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334927 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="registry-server" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334935 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="registry-server" Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334948 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="extract-utilities" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334955 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="extract-utilities" Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334970 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08644b36-c6a1-4c4f-9e4d-89e861f2c197" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.334979 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="08644b36-c6a1-4c4f-9e4d-89e861f2c197" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 06:34:46 crc kubenswrapper[4685]: E1204 06:34:46.334999 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="extract-content" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.335006 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="extract-content" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.335334 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="08644b36-c6a1-4c4f-9e4d-89e861f2c197" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.335348 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe94ddc-d25d-4020-b08c-4e01c6f0055f" containerName="registry-server" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.335365 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc60e06f-e78e-4f53-9e9b-af826e3b8447" containerName="registry-server" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.336081 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.341242 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.341859 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.342296 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.343941 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.354367 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2"] Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.427167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.427239 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.427465 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjftb\" (UniqueName: \"kubernetes.io/projected/398c2009-ba3b-49fa-b49d-838d80fe4bea-kube-api-access-fjftb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.427581 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.529398 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjftb\" (UniqueName: \"kubernetes.io/projected/398c2009-ba3b-49fa-b49d-838d80fe4bea-kube-api-access-fjftb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.529468 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.529540 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.529567 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.534148 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.534530 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.536989 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.553977 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjftb\" (UniqueName: \"kubernetes.io/projected/398c2009-ba3b-49fa-b49d-838d80fe4bea-kube-api-access-fjftb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:46 crc kubenswrapper[4685]: I1204 06:34:46.679377 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:34:47 crc kubenswrapper[4685]: I1204 06:34:47.292605 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2"] Dec 04 06:34:47 crc kubenswrapper[4685]: W1204 06:34:47.293366 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod398c2009_ba3b_49fa_b49d_838d80fe4bea.slice/crio-cd35b4a9fcf439c11bd7f4a34811d7ddb940f08e75320ee0b26598b67bf91bb4 WatchSource:0}: Error finding container cd35b4a9fcf439c11bd7f4a34811d7ddb940f08e75320ee0b26598b67bf91bb4: Status 404 returned error can't find the container with id cd35b4a9fcf439c11bd7f4a34811d7ddb940f08e75320ee0b26598b67bf91bb4 Dec 04 06:34:48 crc kubenswrapper[4685]: I1204 06:34:48.262786 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" event={"ID":"398c2009-ba3b-49fa-b49d-838d80fe4bea","Type":"ContainerStarted","Data":"20849a8bfd2897b5e83186a6a32dc6bcd9212bbfaaa3795bba01315b222db89c"} Dec 04 06:34:48 crc kubenswrapper[4685]: I1204 06:34:48.263123 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" event={"ID":"398c2009-ba3b-49fa-b49d-838d80fe4bea","Type":"ContainerStarted","Data":"cd35b4a9fcf439c11bd7f4a34811d7ddb940f08e75320ee0b26598b67bf91bb4"} Dec 04 06:34:48 crc kubenswrapper[4685]: I1204 06:34:48.292536 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" podStartSLOduration=2.136989504 podStartE2EDuration="2.292515189s" podCreationTimestamp="2025-12-04 06:34:46 +0000 UTC" firstStartedPulling="2025-12-04 06:34:47.296920373 +0000 UTC m=+1424.485151158" lastFinishedPulling="2025-12-04 06:34:47.452446068 +0000 UTC m=+1424.640676843" observedRunningTime="2025-12-04 06:34:48.280531409 +0000 UTC m=+1425.468762204" watchObservedRunningTime="2025-12-04 06:34:48.292515189 +0000 UTC m=+1425.480745974" Dec 04 06:35:15 crc kubenswrapper[4685]: I1204 06:35:15.762439 4685 scope.go:117] "RemoveContainer" containerID="1983666a453ada04c22d70ad58c4de6d7f7153e22fd830178d264146af484ab0" Dec 04 06:35:15 crc kubenswrapper[4685]: I1204 06:35:15.802257 4685 scope.go:117] "RemoveContainer" containerID="c69fbb5aee85e8c9939d6097e3f18da98a2b6f751df6ea1ef0ec34bf5c237bb8" Dec 04 06:35:15 crc kubenswrapper[4685]: I1204 06:35:15.862156 4685 scope.go:117] "RemoveContainer" containerID="9dd67008de7ea673e410a7b4e0869ffe00a03de06e4557b6f96c2f152d815e7c" Dec 04 06:35:15 crc kubenswrapper[4685]: I1204 06:35:15.904799 4685 scope.go:117] "RemoveContainer" containerID="251b7c3911050af23cf8810e5cd4c3046a1a62b7b20c47a86a29e0c37283ccfe" Dec 04 06:35:36 crc kubenswrapper[4685]: I1204 06:35:36.120848 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:35:36 crc kubenswrapper[4685]: I1204 06:35:36.121489 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.487602 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c9cxf"] Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.492079 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.504540 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9cxf"] Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.593912 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-utilities\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.594268 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2424\" (UniqueName: \"kubernetes.io/projected/e5365888-1c35-4996-b592-bc3ed2adda09-kube-api-access-x2424\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.594603 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-catalog-content\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.695960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-utilities\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.696064 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2424\" (UniqueName: \"kubernetes.io/projected/e5365888-1c35-4996-b592-bc3ed2adda09-kube-api-access-x2424\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.696101 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-catalog-content\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.696672 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-catalog-content\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.696882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-utilities\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.719639 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2424\" (UniqueName: \"kubernetes.io/projected/e5365888-1c35-4996-b592-bc3ed2adda09-kube-api-access-x2424\") pod \"community-operators-c9cxf\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:01 crc kubenswrapper[4685]: I1204 06:36:01.828042 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:02 crc kubenswrapper[4685]: I1204 06:36:02.344986 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9cxf"] Dec 04 06:36:03 crc kubenswrapper[4685]: I1204 06:36:03.224319 4685 generic.go:334] "Generic (PLEG): container finished" podID="e5365888-1c35-4996-b592-bc3ed2adda09" containerID="0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6" exitCode=0 Dec 04 06:36:03 crc kubenswrapper[4685]: I1204 06:36:03.224393 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerDied","Data":"0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6"} Dec 04 06:36:03 crc kubenswrapper[4685]: I1204 06:36:03.224794 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerStarted","Data":"84a9cbf4f38309c1be22c8c72f8c1bba538b4747088a6814afd8927e8b1a396a"} Dec 04 06:36:04 crc kubenswrapper[4685]: I1204 06:36:04.260116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerStarted","Data":"0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f"} Dec 04 06:36:05 crc kubenswrapper[4685]: I1204 06:36:05.278118 4685 generic.go:334] "Generic (PLEG): container finished" podID="e5365888-1c35-4996-b592-bc3ed2adda09" containerID="0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f" exitCode=0 Dec 04 06:36:05 crc kubenswrapper[4685]: I1204 06:36:05.278179 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerDied","Data":"0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f"} Dec 04 06:36:06 crc kubenswrapper[4685]: I1204 06:36:06.122055 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:36:06 crc kubenswrapper[4685]: I1204 06:36:06.122131 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:36:07 crc kubenswrapper[4685]: I1204 06:36:07.302262 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerStarted","Data":"8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a"} Dec 04 06:36:07 crc kubenswrapper[4685]: I1204 06:36:07.330254 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c9cxf" podStartSLOduration=3.311118636 podStartE2EDuration="6.330221182s" podCreationTimestamp="2025-12-04 06:36:01 +0000 UTC" firstStartedPulling="2025-12-04 06:36:03.226669556 +0000 UTC m=+1500.414900321" lastFinishedPulling="2025-12-04 06:36:06.245772092 +0000 UTC m=+1503.434002867" observedRunningTime="2025-12-04 06:36:07.321096911 +0000 UTC m=+1504.509327736" watchObservedRunningTime="2025-12-04 06:36:07.330221182 +0000 UTC m=+1504.518452007" Dec 04 06:36:11 crc kubenswrapper[4685]: I1204 06:36:11.829122 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:11 crc kubenswrapper[4685]: I1204 06:36:11.829796 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:11 crc kubenswrapper[4685]: I1204 06:36:11.906096 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:12 crc kubenswrapper[4685]: I1204 06:36:12.406700 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:12 crc kubenswrapper[4685]: I1204 06:36:12.457229 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9cxf"] Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.372398 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c9cxf" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="registry-server" containerID="cri-o://8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a" gracePeriod=2 Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.839392 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.877548 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2424\" (UniqueName: \"kubernetes.io/projected/e5365888-1c35-4996-b592-bc3ed2adda09-kube-api-access-x2424\") pod \"e5365888-1c35-4996-b592-bc3ed2adda09\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.877726 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-utilities\") pod \"e5365888-1c35-4996-b592-bc3ed2adda09\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.877844 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-catalog-content\") pod \"e5365888-1c35-4996-b592-bc3ed2adda09\" (UID: \"e5365888-1c35-4996-b592-bc3ed2adda09\") " Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.879263 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-utilities" (OuterVolumeSpecName: "utilities") pod "e5365888-1c35-4996-b592-bc3ed2adda09" (UID: "e5365888-1c35-4996-b592-bc3ed2adda09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.883467 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5365888-1c35-4996-b592-bc3ed2adda09-kube-api-access-x2424" (OuterVolumeSpecName: "kube-api-access-x2424") pod "e5365888-1c35-4996-b592-bc3ed2adda09" (UID: "e5365888-1c35-4996-b592-bc3ed2adda09"). InnerVolumeSpecName "kube-api-access-x2424". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.980633 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2424\" (UniqueName: \"kubernetes.io/projected/e5365888-1c35-4996-b592-bc3ed2adda09-kube-api-access-x2424\") on node \"crc\" DevicePath \"\"" Dec 04 06:36:14 crc kubenswrapper[4685]: I1204 06:36:14.980874 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.111980 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5365888-1c35-4996-b592-bc3ed2adda09" (UID: "e5365888-1c35-4996-b592-bc3ed2adda09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.183934 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5365888-1c35-4996-b592-bc3ed2adda09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.389390 4685 generic.go:334] "Generic (PLEG): container finished" podID="e5365888-1c35-4996-b592-bc3ed2adda09" containerID="8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a" exitCode=0 Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.389518 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9cxf" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.389477 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerDied","Data":"8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a"} Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.389709 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9cxf" event={"ID":"e5365888-1c35-4996-b592-bc3ed2adda09","Type":"ContainerDied","Data":"84a9cbf4f38309c1be22c8c72f8c1bba538b4747088a6814afd8927e8b1a396a"} Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.389755 4685 scope.go:117] "RemoveContainer" containerID="8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.417039 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9cxf"] Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.426357 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c9cxf"] Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.445935 4685 scope.go:117] "RemoveContainer" containerID="0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.483352 4685 scope.go:117] "RemoveContainer" containerID="0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.521603 4685 scope.go:117] "RemoveContainer" containerID="8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a" Dec 04 06:36:15 crc kubenswrapper[4685]: E1204 06:36:15.522144 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a\": container with ID starting with 8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a not found: ID does not exist" containerID="8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.522188 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a"} err="failed to get container status \"8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a\": rpc error: code = NotFound desc = could not find container \"8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a\": container with ID starting with 8dfc9da7a4316eab11b1ccb08410bf4036ce7f152502da0623320ef3c999d28a not found: ID does not exist" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.522216 4685 scope.go:117] "RemoveContainer" containerID="0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f" Dec 04 06:36:15 crc kubenswrapper[4685]: E1204 06:36:15.522641 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f\": container with ID starting with 0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f not found: ID does not exist" containerID="0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.522703 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f"} err="failed to get container status \"0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f\": rpc error: code = NotFound desc = could not find container \"0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f\": container with ID starting with 0b02223d60088b5866890a327b1f59802f6ca66c07017f6e7fbc44a2bfcf6b9f not found: ID does not exist" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.522743 4685 scope.go:117] "RemoveContainer" containerID="0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6" Dec 04 06:36:15 crc kubenswrapper[4685]: E1204 06:36:15.523161 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6\": container with ID starting with 0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6 not found: ID does not exist" containerID="0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6" Dec 04 06:36:15 crc kubenswrapper[4685]: I1204 06:36:15.523181 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6"} err="failed to get container status \"0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6\": rpc error: code = NotFound desc = could not find container \"0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6\": container with ID starting with 0da003a4f2f0470a12a05876ff513d1592aceae19316bc6c716d25d2b50e86f6 not found: ID does not exist" Dec 04 06:36:16 crc kubenswrapper[4685]: I1204 06:36:16.047837 4685 scope.go:117] "RemoveContainer" containerID="5e32e6824d22ae1aa3272b6cd47136721ab3ad5274f2f968f1bc3abb29d018c5" Dec 04 06:36:16 crc kubenswrapper[4685]: I1204 06:36:16.096605 4685 scope.go:117] "RemoveContainer" containerID="7195b2be6943bd0983daf635bd5ec023782817786ac253eb69ce2b406968b974" Dec 04 06:36:16 crc kubenswrapper[4685]: I1204 06:36:16.384394 4685 scope.go:117] "RemoveContainer" containerID="b7effbed9be6c6720b522757019ea495f31b64e5e9664c9384e58554e62eff9c" Dec 04 06:36:16 crc kubenswrapper[4685]: I1204 06:36:16.411474 4685 scope.go:117] "RemoveContainer" containerID="252f6828706c15e58fcb7a3b491ecee0a37c85fd571fbefe50fcd5492a05e612" Dec 04 06:36:16 crc kubenswrapper[4685]: I1204 06:36:16.473872 4685 scope.go:117] "RemoveContainer" containerID="179ee4c4f3bb1d39611f2e3d12131ac9da868bf495deca3146a5b9cc40bd0ed6" Dec 04 06:36:17 crc kubenswrapper[4685]: I1204 06:36:17.144616 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" path="/var/lib/kubelet/pods/e5365888-1c35-4996-b592-bc3ed2adda09/volumes" Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.121311 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.121902 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.121951 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.122602 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.122666 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" gracePeriod=600 Dec 04 06:36:36 crc kubenswrapper[4685]: E1204 06:36:36.258703 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.645572 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" exitCode=0 Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.645624 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47"} Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.645666 4685 scope.go:117] "RemoveContainer" containerID="6fc1b67a51a8fb7e262f18c32ce16b7ebc33919ea2c7327f8e192aec0ec81fc6" Dec 04 06:36:36 crc kubenswrapper[4685]: I1204 06:36:36.646481 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:36:36 crc kubenswrapper[4685]: E1204 06:36:36.646971 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.988578 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lz945"] Dec 04 06:36:37 crc kubenswrapper[4685]: E1204 06:36:37.989323 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="extract-content" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.989339 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="extract-content" Dec 04 06:36:37 crc kubenswrapper[4685]: E1204 06:36:37.989359 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="extract-utilities" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.989367 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="extract-utilities" Dec 04 06:36:37 crc kubenswrapper[4685]: E1204 06:36:37.989423 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="registry-server" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.989429 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="registry-server" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.989599 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5365888-1c35-4996-b592-bc3ed2adda09" containerName="registry-server" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.991020 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:37 crc kubenswrapper[4685]: I1204 06:36:37.997437 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lz945"] Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.108209 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-catalog-content\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.108824 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-utilities\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.108940 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv44z\" (UniqueName: \"kubernetes.io/projected/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-kube-api-access-sv44z\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.211317 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-catalog-content\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.211384 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-utilities\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.211530 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv44z\" (UniqueName: \"kubernetes.io/projected/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-kube-api-access-sv44z\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.212686 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-utilities\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.212717 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-catalog-content\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.237011 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv44z\" (UniqueName: \"kubernetes.io/projected/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-kube-api-access-sv44z\") pod \"certified-operators-lz945\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.312040 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:38 crc kubenswrapper[4685]: I1204 06:36:38.800426 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lz945"] Dec 04 06:36:39 crc kubenswrapper[4685]: I1204 06:36:39.679135 4685 generic.go:334] "Generic (PLEG): container finished" podID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerID="d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae" exitCode=0 Dec 04 06:36:39 crc kubenswrapper[4685]: I1204 06:36:39.679173 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz945" event={"ID":"ad4e0815-c1a3-48e4-a365-0db1010f7bc1","Type":"ContainerDied","Data":"d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae"} Dec 04 06:36:39 crc kubenswrapper[4685]: I1204 06:36:39.679196 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz945" event={"ID":"ad4e0815-c1a3-48e4-a365-0db1010f7bc1","Type":"ContainerStarted","Data":"ac1405363a794344e594271431adb1e6b65fe5ac969ea2de9affd0a083a9568f"} Dec 04 06:36:41 crc kubenswrapper[4685]: I1204 06:36:41.703343 4685 generic.go:334] "Generic (PLEG): container finished" podID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerID="c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366" exitCode=0 Dec 04 06:36:41 crc kubenswrapper[4685]: I1204 06:36:41.703477 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz945" event={"ID":"ad4e0815-c1a3-48e4-a365-0db1010f7bc1","Type":"ContainerDied","Data":"c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366"} Dec 04 06:36:42 crc kubenswrapper[4685]: I1204 06:36:42.717051 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz945" event={"ID":"ad4e0815-c1a3-48e4-a365-0db1010f7bc1","Type":"ContainerStarted","Data":"01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28"} Dec 04 06:36:42 crc kubenswrapper[4685]: I1204 06:36:42.754270 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lz945" podStartSLOduration=3.282842596 podStartE2EDuration="5.754243042s" podCreationTimestamp="2025-12-04 06:36:37 +0000 UTC" firstStartedPulling="2025-12-04 06:36:39.681028105 +0000 UTC m=+1536.869258880" lastFinishedPulling="2025-12-04 06:36:42.152428541 +0000 UTC m=+1539.340659326" observedRunningTime="2025-12-04 06:36:42.73995382 +0000 UTC m=+1539.928184595" watchObservedRunningTime="2025-12-04 06:36:42.754243042 +0000 UTC m=+1539.942473817" Dec 04 06:36:48 crc kubenswrapper[4685]: I1204 06:36:48.312714 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:48 crc kubenswrapper[4685]: I1204 06:36:48.313222 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:48 crc kubenswrapper[4685]: I1204 06:36:48.365279 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:48 crc kubenswrapper[4685]: I1204 06:36:48.867558 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:48 crc kubenswrapper[4685]: I1204 06:36:48.941850 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lz945"] Dec 04 06:36:49 crc kubenswrapper[4685]: I1204 06:36:49.125801 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:36:49 crc kubenswrapper[4685]: E1204 06:36:49.126229 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:36:50 crc kubenswrapper[4685]: I1204 06:36:50.810211 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lz945" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="registry-server" containerID="cri-o://01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28" gracePeriod=2 Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.295042 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.412463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-utilities\") pod \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.412599 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv44z\" (UniqueName: \"kubernetes.io/projected/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-kube-api-access-sv44z\") pod \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.412698 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-catalog-content\") pod \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\" (UID: \"ad4e0815-c1a3-48e4-a365-0db1010f7bc1\") " Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.413140 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-utilities" (OuterVolumeSpecName: "utilities") pod "ad4e0815-c1a3-48e4-a365-0db1010f7bc1" (UID: "ad4e0815-c1a3-48e4-a365-0db1010f7bc1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.418871 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-kube-api-access-sv44z" (OuterVolumeSpecName: "kube-api-access-sv44z") pod "ad4e0815-c1a3-48e4-a365-0db1010f7bc1" (UID: "ad4e0815-c1a3-48e4-a365-0db1010f7bc1"). InnerVolumeSpecName "kube-api-access-sv44z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.514278 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv44z\" (UniqueName: \"kubernetes.io/projected/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-kube-api-access-sv44z\") on node \"crc\" DevicePath \"\"" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.514304 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.533617 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad4e0815-c1a3-48e4-a365-0db1010f7bc1" (UID: "ad4e0815-c1a3-48e4-a365-0db1010f7bc1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.615474 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4e0815-c1a3-48e4-a365-0db1010f7bc1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.820866 4685 generic.go:334] "Generic (PLEG): container finished" podID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerID="01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28" exitCode=0 Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.820921 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz945" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.820949 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz945" event={"ID":"ad4e0815-c1a3-48e4-a365-0db1010f7bc1","Type":"ContainerDied","Data":"01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28"} Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.821011 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz945" event={"ID":"ad4e0815-c1a3-48e4-a365-0db1010f7bc1","Type":"ContainerDied","Data":"ac1405363a794344e594271431adb1e6b65fe5ac969ea2de9affd0a083a9568f"} Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.821032 4685 scope.go:117] "RemoveContainer" containerID="01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.849853 4685 scope.go:117] "RemoveContainer" containerID="c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.859115 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lz945"] Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.869459 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lz945"] Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.874893 4685 scope.go:117] "RemoveContainer" containerID="d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.920634 4685 scope.go:117] "RemoveContainer" containerID="01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28" Dec 04 06:36:51 crc kubenswrapper[4685]: E1204 06:36:51.920968 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28\": container with ID starting with 01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28 not found: ID does not exist" containerID="01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.920998 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28"} err="failed to get container status \"01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28\": rpc error: code = NotFound desc = could not find container \"01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28\": container with ID starting with 01bf9ede83ab9abdb0c263d098c1ad6923413d0c5126ae18db7eae70bacfbb28 not found: ID does not exist" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.921019 4685 scope.go:117] "RemoveContainer" containerID="c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366" Dec 04 06:36:51 crc kubenswrapper[4685]: E1204 06:36:51.921185 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366\": container with ID starting with c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366 not found: ID does not exist" containerID="c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.921208 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366"} err="failed to get container status \"c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366\": rpc error: code = NotFound desc = could not find container \"c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366\": container with ID starting with c84bf240ccd847ba641edf2e46314cb04aedccf7d930412076db473e857c3366 not found: ID does not exist" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.921221 4685 scope.go:117] "RemoveContainer" containerID="d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae" Dec 04 06:36:51 crc kubenswrapper[4685]: E1204 06:36:51.921503 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae\": container with ID starting with d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae not found: ID does not exist" containerID="d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae" Dec 04 06:36:51 crc kubenswrapper[4685]: I1204 06:36:51.921522 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae"} err="failed to get container status \"d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae\": rpc error: code = NotFound desc = could not find container \"d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae\": container with ID starting with d393dd2be9b07447e7660d93810a245cfa1b76f8948d99b5c620df4654e9aaae not found: ID does not exist" Dec 04 06:36:53 crc kubenswrapper[4685]: I1204 06:36:53.147910 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" path="/var/lib/kubelet/pods/ad4e0815-c1a3-48e4-a365-0db1010f7bc1/volumes" Dec 04 06:37:00 crc kubenswrapper[4685]: I1204 06:37:00.126254 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:37:00 crc kubenswrapper[4685]: E1204 06:37:00.127179 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:37:14 crc kubenswrapper[4685]: I1204 06:37:14.126608 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:37:14 crc kubenswrapper[4685]: E1204 06:37:14.127577 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:37:25 crc kubenswrapper[4685]: I1204 06:37:25.126365 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:37:25 crc kubenswrapper[4685]: E1204 06:37:25.127348 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:37:38 crc kubenswrapper[4685]: I1204 06:37:38.127729 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:37:38 crc kubenswrapper[4685]: E1204 06:37:38.128602 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:37:49 crc kubenswrapper[4685]: I1204 06:37:49.126156 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:37:49 crc kubenswrapper[4685]: E1204 06:37:49.129511 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:38:00 crc kubenswrapper[4685]: I1204 06:38:00.633535 4685 generic.go:334] "Generic (PLEG): container finished" podID="398c2009-ba3b-49fa-b49d-838d80fe4bea" containerID="20849a8bfd2897b5e83186a6a32dc6bcd9212bbfaaa3795bba01315b222db89c" exitCode=0 Dec 04 06:38:00 crc kubenswrapper[4685]: I1204 06:38:00.633613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" event={"ID":"398c2009-ba3b-49fa-b49d-838d80fe4bea","Type":"ContainerDied","Data":"20849a8bfd2897b5e83186a6a32dc6bcd9212bbfaaa3795bba01315b222db89c"} Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.017617 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.097636 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-ssh-key\") pod \"398c2009-ba3b-49fa-b49d-838d80fe4bea\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.097699 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-bootstrap-combined-ca-bundle\") pod \"398c2009-ba3b-49fa-b49d-838d80fe4bea\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.097721 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjftb\" (UniqueName: \"kubernetes.io/projected/398c2009-ba3b-49fa-b49d-838d80fe4bea-kube-api-access-fjftb\") pod \"398c2009-ba3b-49fa-b49d-838d80fe4bea\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.097824 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-inventory\") pod \"398c2009-ba3b-49fa-b49d-838d80fe4bea\" (UID: \"398c2009-ba3b-49fa-b49d-838d80fe4bea\") " Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.104660 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/398c2009-ba3b-49fa-b49d-838d80fe4bea-kube-api-access-fjftb" (OuterVolumeSpecName: "kube-api-access-fjftb") pod "398c2009-ba3b-49fa-b49d-838d80fe4bea" (UID: "398c2009-ba3b-49fa-b49d-838d80fe4bea"). InnerVolumeSpecName "kube-api-access-fjftb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.105493 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "398c2009-ba3b-49fa-b49d-838d80fe4bea" (UID: "398c2009-ba3b-49fa-b49d-838d80fe4bea"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.125422 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "398c2009-ba3b-49fa-b49d-838d80fe4bea" (UID: "398c2009-ba3b-49fa-b49d-838d80fe4bea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.126123 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:38:02 crc kubenswrapper[4685]: E1204 06:38:02.126477 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.150275 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-inventory" (OuterVolumeSpecName: "inventory") pod "398c2009-ba3b-49fa-b49d-838d80fe4bea" (UID: "398c2009-ba3b-49fa-b49d-838d80fe4bea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.200542 4685 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.200572 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjftb\" (UniqueName: \"kubernetes.io/projected/398c2009-ba3b-49fa-b49d-838d80fe4bea-kube-api-access-fjftb\") on node \"crc\" DevicePath \"\"" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.200583 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.200591 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/398c2009-ba3b-49fa-b49d-838d80fe4bea-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.662816 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" event={"ID":"398c2009-ba3b-49fa-b49d-838d80fe4bea","Type":"ContainerDied","Data":"cd35b4a9fcf439c11bd7f4a34811d7ddb940f08e75320ee0b26598b67bf91bb4"} Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.662871 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.662893 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd35b4a9fcf439c11bd7f4a34811d7ddb940f08e75320ee0b26598b67bf91bb4" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.764941 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w"] Dec 04 06:38:02 crc kubenswrapper[4685]: E1204 06:38:02.765466 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="extract-content" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.765486 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="extract-content" Dec 04 06:38:02 crc kubenswrapper[4685]: E1204 06:38:02.765525 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="registry-server" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.765534 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="registry-server" Dec 04 06:38:02 crc kubenswrapper[4685]: E1204 06:38:02.765565 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398c2009-ba3b-49fa-b49d-838d80fe4bea" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.765575 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="398c2009-ba3b-49fa-b49d-838d80fe4bea" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 06:38:02 crc kubenswrapper[4685]: E1204 06:38:02.765593 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="extract-utilities" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.765601 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="extract-utilities" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.765843 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad4e0815-c1a3-48e4-a365-0db1010f7bc1" containerName="registry-server" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.765875 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="398c2009-ba3b-49fa-b49d-838d80fe4bea" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.766708 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.768894 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.769018 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.769238 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.769461 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.777976 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w"] Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.821956 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.822074 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.822283 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5cmh\" (UniqueName: \"kubernetes.io/projected/701d7b10-8594-480e-8b64-ee2f6513b970-kube-api-access-k5cmh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.924134 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5cmh\" (UniqueName: \"kubernetes.io/projected/701d7b10-8594-480e-8b64-ee2f6513b970-kube-api-access-k5cmh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.924679 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.924738 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.928501 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.929864 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:02 crc kubenswrapper[4685]: I1204 06:38:02.944364 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5cmh\" (UniqueName: \"kubernetes.io/projected/701d7b10-8594-480e-8b64-ee2f6513b970-kube-api-access-k5cmh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-97x4w\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:03 crc kubenswrapper[4685]: I1204 06:38:03.142297 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:38:03 crc kubenswrapper[4685]: I1204 06:38:03.150336 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:38:03 crc kubenswrapper[4685]: I1204 06:38:03.703449 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w"] Dec 04 06:38:03 crc kubenswrapper[4685]: W1204 06:38:03.710222 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod701d7b10_8594_480e_8b64_ee2f6513b970.slice/crio-1d0789965461bfc41bed515032c7846028a9579b7e33959fb0ddd542a983bf75 WatchSource:0}: Error finding container 1d0789965461bfc41bed515032c7846028a9579b7e33959fb0ddd542a983bf75: Status 404 returned error can't find the container with id 1d0789965461bfc41bed515032c7846028a9579b7e33959fb0ddd542a983bf75 Dec 04 06:38:03 crc kubenswrapper[4685]: I1204 06:38:03.929525 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:38:04 crc kubenswrapper[4685]: I1204 06:38:04.685676 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" event={"ID":"701d7b10-8594-480e-8b64-ee2f6513b970","Type":"ContainerStarted","Data":"02c481755d1fff5123ed2584c6adf9503c01f65c8a0adc70863b5b08ece7b4f0"} Dec 04 06:38:04 crc kubenswrapper[4685]: I1204 06:38:04.685954 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" event={"ID":"701d7b10-8594-480e-8b64-ee2f6513b970","Type":"ContainerStarted","Data":"1d0789965461bfc41bed515032c7846028a9579b7e33959fb0ddd542a983bf75"} Dec 04 06:38:15 crc kubenswrapper[4685]: I1204 06:38:15.126155 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:38:15 crc kubenswrapper[4685]: E1204 06:38:15.127139 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:38:16 crc kubenswrapper[4685]: I1204 06:38:16.757901 4685 scope.go:117] "RemoveContainer" containerID="037bcd5d9c5e1a269be02ec8f9a987e72891d759a570d0b8235167400f87a243" Dec 04 06:38:16 crc kubenswrapper[4685]: I1204 06:38:16.795315 4685 scope.go:117] "RemoveContainer" containerID="bc4b1131ba7c64eacad612b27d0cde155fe0fadffa3bf7d74fe94635678af5d1" Dec 04 06:38:28 crc kubenswrapper[4685]: I1204 06:38:28.126492 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:38:28 crc kubenswrapper[4685]: E1204 06:38:28.127868 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:38:43 crc kubenswrapper[4685]: I1204 06:38:43.140147 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:38:43 crc kubenswrapper[4685]: E1204 06:38:43.141071 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.053117 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" podStartSLOduration=52.839289666 podStartE2EDuration="53.053093412s" podCreationTimestamp="2025-12-04 06:38:02 +0000 UTC" firstStartedPulling="2025-12-04 06:38:03.712498917 +0000 UTC m=+1620.900729702" lastFinishedPulling="2025-12-04 06:38:03.926302673 +0000 UTC m=+1621.114533448" observedRunningTime="2025-12-04 06:38:04.712935868 +0000 UTC m=+1621.901166673" watchObservedRunningTime="2025-12-04 06:38:55.053093412 +0000 UTC m=+1672.241324207" Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.056800 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jn7p4"] Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.064863 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-kfvw5"] Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.074360 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cd77-account-create-update-2vkcr"] Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.084804 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cd77-account-create-update-2vkcr"] Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.093431 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jn7p4"] Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.104797 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-kfvw5"] Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.142099 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf58941-8b22-419f-bcd6-27bd436ddde4" path="/var/lib/kubelet/pods/4cf58941-8b22-419f-bcd6-27bd436ddde4/volumes" Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.143281 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9214ce67-d976-4159-9a3b-dcc8d5ff3c1d" path="/var/lib/kubelet/pods/9214ce67-d976-4159-9a3b-dcc8d5ff3c1d/volumes" Dec 04 06:38:55 crc kubenswrapper[4685]: I1204 06:38:55.144717 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e844c8fb-db9a-4bc4-a0a1-4fe116f05608" path="/var/lib/kubelet/pods/e844c8fb-db9a-4bc4-a0a1-4fe116f05608/volumes" Dec 04 06:38:56 crc kubenswrapper[4685]: I1204 06:38:56.037894 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9056-account-create-update-2sqd8"] Dec 04 06:38:56 crc kubenswrapper[4685]: I1204 06:38:56.050689 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dsnhj"] Dec 04 06:38:56 crc kubenswrapper[4685]: I1204 06:38:56.064453 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9056-account-create-update-2sqd8"] Dec 04 06:38:56 crc kubenswrapper[4685]: I1204 06:38:56.073671 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8a6e-account-create-update-vhvkz"] Dec 04 06:38:56 crc kubenswrapper[4685]: I1204 06:38:56.082045 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dsnhj"] Dec 04 06:38:56 crc kubenswrapper[4685]: I1204 06:38:56.090073 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8a6e-account-create-update-vhvkz"] Dec 04 06:38:57 crc kubenswrapper[4685]: I1204 06:38:57.127153 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:38:57 crc kubenswrapper[4685]: E1204 06:38:57.127600 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:38:57 crc kubenswrapper[4685]: I1204 06:38:57.140732 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6edc5c45-5581-4a04-9414-847a4c817439" path="/var/lib/kubelet/pods/6edc5c45-5581-4a04-9414-847a4c817439/volumes" Dec 04 06:38:57 crc kubenswrapper[4685]: I1204 06:38:57.141534 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78799f6c-2669-4a97-b50d-a772ee9b7dea" path="/var/lib/kubelet/pods/78799f6c-2669-4a97-b50d-a772ee9b7dea/volumes" Dec 04 06:38:57 crc kubenswrapper[4685]: I1204 06:38:57.142318 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab4e10f7-15cc-42da-9a4b-80738fb81cb2" path="/var/lib/kubelet/pods/ab4e10f7-15cc-42da-9a4b-80738fb81cb2/volumes" Dec 04 06:39:12 crc kubenswrapper[4685]: I1204 06:39:12.126250 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:39:12 crc kubenswrapper[4685]: E1204 06:39:12.126976 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:39:16 crc kubenswrapper[4685]: I1204 06:39:16.863380 4685 scope.go:117] "RemoveContainer" containerID="d50bbb1bb2816fc757797cccc70048dd6e28dd963f4e1b16296b8ed76e964693" Dec 04 06:39:16 crc kubenswrapper[4685]: I1204 06:39:16.899741 4685 scope.go:117] "RemoveContainer" containerID="3dc417c7088031ac631b07ca8f46de81e93ee3587153b26c3f906708394d9e5b" Dec 04 06:39:16 crc kubenswrapper[4685]: I1204 06:39:16.983311 4685 scope.go:117] "RemoveContainer" containerID="dba855c2804e36e3af395966a469b7b9eac03d6d2f0243dcc8253af7ed252f1d" Dec 04 06:39:17 crc kubenswrapper[4685]: I1204 06:39:17.037917 4685 scope.go:117] "RemoveContainer" containerID="7f80d8af48f1348c02c77eb50dbdf263cd1ace923a39ca6358f662f61a1a4cb4" Dec 04 06:39:17 crc kubenswrapper[4685]: I1204 06:39:17.082368 4685 scope.go:117] "RemoveContainer" containerID="c43e0b227c96c5ec640cb727d5443cab2edb0ee417eaffa427af003de0a8f88f" Dec 04 06:39:17 crc kubenswrapper[4685]: I1204 06:39:17.146139 4685 scope.go:117] "RemoveContainer" containerID="e67ca61666265757fdc38ab13f318dea4d38ca0f1be22bc8338befdeec1a7deb" Dec 04 06:39:19 crc kubenswrapper[4685]: I1204 06:39:19.078011 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8bkmd"] Dec 04 06:39:19 crc kubenswrapper[4685]: I1204 06:39:19.093097 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8bkmd"] Dec 04 06:39:19 crc kubenswrapper[4685]: I1204 06:39:19.137169 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0" path="/var/lib/kubelet/pods/2733d0b6-b52b-4e2e-836b-5dd09dc7b2f0/volumes" Dec 04 06:39:23 crc kubenswrapper[4685]: I1204 06:39:23.138668 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:39:23 crc kubenswrapper[4685]: E1204 06:39:23.139324 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:39:36 crc kubenswrapper[4685]: I1204 06:39:36.125246 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:39:36 crc kubenswrapper[4685]: E1204 06:39:36.125908 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.074601 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5b79-account-create-update-bdgk2"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.086830 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-41f6-account-create-update-ctjs8"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.098003 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-sh2ng"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.109993 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tmptx"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.119281 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5b79-account-create-update-bdgk2"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.128625 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-f5f0-account-create-update-gl7hp"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.136045 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tmptx"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.143839 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-v6sh8"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.151320 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-sh2ng"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.158763 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-f5f0-account-create-update-gl7hp"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.165537 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-41f6-account-create-update-ctjs8"] Dec 04 06:39:38 crc kubenswrapper[4685]: I1204 06:39:38.172053 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-v6sh8"] Dec 04 06:39:39 crc kubenswrapper[4685]: I1204 06:39:39.142129 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a0c2c44-843d-4750-abbe-b4a9effe469f" path="/var/lib/kubelet/pods/3a0c2c44-843d-4750-abbe-b4a9effe469f/volumes" Dec 04 06:39:39 crc kubenswrapper[4685]: I1204 06:39:39.144020 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="989f07ae-40ac-4b8d-91de-167df369b845" path="/var/lib/kubelet/pods/989f07ae-40ac-4b8d-91de-167df369b845/volumes" Dec 04 06:39:39 crc kubenswrapper[4685]: I1204 06:39:39.145365 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b071b469-ce94-4165-a933-55f9f8f36abe" path="/var/lib/kubelet/pods/b071b469-ce94-4165-a933-55f9f8f36abe/volumes" Dec 04 06:39:39 crc kubenswrapper[4685]: I1204 06:39:39.146861 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfbaf5e7-e351-4c00-83e1-a3131589e9d9" path="/var/lib/kubelet/pods/bfbaf5e7-e351-4c00-83e1-a3131589e9d9/volumes" Dec 04 06:39:39 crc kubenswrapper[4685]: I1204 06:39:39.152394 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dabcb5c7-61d1-46eb-9c15-c14a1e0df79e" path="/var/lib/kubelet/pods/dabcb5c7-61d1-46eb-9c15-c14a1e0df79e/volumes" Dec 04 06:39:39 crc kubenswrapper[4685]: I1204 06:39:39.154473 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea062396-352c-4a83-8745-5bec1a9284bc" path="/var/lib/kubelet/pods/ea062396-352c-4a83-8745-5bec1a9284bc/volumes" Dec 04 06:39:42 crc kubenswrapper[4685]: I1204 06:39:42.027631 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wlwr5"] Dec 04 06:39:42 crc kubenswrapper[4685]: I1204 06:39:42.038664 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wlwr5"] Dec 04 06:39:43 crc kubenswrapper[4685]: I1204 06:39:43.141140 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd467839-0575-42ac-98fd-4e6d2cb3ceba" path="/var/lib/kubelet/pods/bd467839-0575-42ac-98fd-4e6d2cb3ceba/volumes" Dec 04 06:39:51 crc kubenswrapper[4685]: I1204 06:39:51.126469 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:39:51 crc kubenswrapper[4685]: E1204 06:39:51.127270 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:40:05 crc kubenswrapper[4685]: I1204 06:40:05.126193 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:40:05 crc kubenswrapper[4685]: E1204 06:40:05.127097 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:40:11 crc kubenswrapper[4685]: I1204 06:40:11.053092 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-bcqrm"] Dec 04 06:40:11 crc kubenswrapper[4685]: I1204 06:40:11.065659 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-bcqrm"] Dec 04 06:40:11 crc kubenswrapper[4685]: I1204 06:40:11.137121 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab9f58d-e6bb-445d-a748-0059a8981f49" path="/var/lib/kubelet/pods/4ab9f58d-e6bb-445d-a748-0059a8981f49/volumes" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.126771 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:40:17 crc kubenswrapper[4685]: E1204 06:40:17.128271 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.286076 4685 scope.go:117] "RemoveContainer" containerID="7343eab1c44d1c3d6cfb8caa591452bbf50e8639bce0115f5b2bb48413586cf9" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.320396 4685 scope.go:117] "RemoveContainer" containerID="9472458b1f7b017c0ddb1d80493683f3e4e415e4cdc6387da6cd9b1b21bd7729" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.410857 4685 scope.go:117] "RemoveContainer" containerID="4dd65d877c9d947c1e27c08f400d71a542c2120ace1d3794a568be83e96f8e79" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.437555 4685 scope.go:117] "RemoveContainer" containerID="608a3a642b882aef1cb59d4b14f807074834a50894a292a70d5a1860a04f9b6a" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.516920 4685 scope.go:117] "RemoveContainer" containerID="0489c7258dea7c5d4594598701151954700bf5703a38d3a569a79c1ec24b8643" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.555053 4685 scope.go:117] "RemoveContainer" containerID="f32dd0f8de4ed06f6817844bfac86477a312740b57b1629d864513823a311230" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.599580 4685 scope.go:117] "RemoveContainer" containerID="93bd3f5f888caee883654231951f728342202843c20f1d659213346365251944" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.622190 4685 scope.go:117] "RemoveContainer" containerID="0b5ac81183e24c805c82016dcc0273fe2f7185d64132e5578c355a40f2ef4ab0" Dec 04 06:40:17 crc kubenswrapper[4685]: I1204 06:40:17.658189 4685 scope.go:117] "RemoveContainer" containerID="6fd7328ae0af3c59a457ba4a7e667dfc305438d00100afc19cf6fa4b2bc15cb0" Dec 04 06:40:23 crc kubenswrapper[4685]: I1204 06:40:23.068030 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bljbv"] Dec 04 06:40:23 crc kubenswrapper[4685]: I1204 06:40:23.089863 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cpkd7"] Dec 04 06:40:23 crc kubenswrapper[4685]: I1204 06:40:23.099848 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bljbv"] Dec 04 06:40:23 crc kubenswrapper[4685]: I1204 06:40:23.108328 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cpkd7"] Dec 04 06:40:23 crc kubenswrapper[4685]: I1204 06:40:23.151174 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3" path="/var/lib/kubelet/pods/6c76ec1c-6d8c-48e0-9128-4fd9fb3cbed3/volumes" Dec 04 06:40:23 crc kubenswrapper[4685]: I1204 06:40:23.152033 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd072a3b-1456-43ad-97a7-7326a66a6f9d" path="/var/lib/kubelet/pods/fd072a3b-1456-43ad-97a7-7326a66a6f9d/volumes" Dec 04 06:40:31 crc kubenswrapper[4685]: I1204 06:40:31.035218 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5498d"] Dec 04 06:40:31 crc kubenswrapper[4685]: I1204 06:40:31.044637 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5498d"] Dec 04 06:40:31 crc kubenswrapper[4685]: I1204 06:40:31.126271 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:40:31 crc kubenswrapper[4685]: E1204 06:40:31.126581 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:40:31 crc kubenswrapper[4685]: I1204 06:40:31.140693 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f6bce91-0f30-4543-ac33-b34877206b05" path="/var/lib/kubelet/pods/6f6bce91-0f30-4543-ac33-b34877206b05/volumes" Dec 04 06:40:41 crc kubenswrapper[4685]: I1204 06:40:41.483621 4685 generic.go:334] "Generic (PLEG): container finished" podID="701d7b10-8594-480e-8b64-ee2f6513b970" containerID="02c481755d1fff5123ed2584c6adf9503c01f65c8a0adc70863b5b08ece7b4f0" exitCode=0 Dec 04 06:40:41 crc kubenswrapper[4685]: I1204 06:40:41.483736 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" event={"ID":"701d7b10-8594-480e-8b64-ee2f6513b970","Type":"ContainerDied","Data":"02c481755d1fff5123ed2584c6adf9503c01f65c8a0adc70863b5b08ece7b4f0"} Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.072500 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-59cdh"] Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.087910 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-59cdh"] Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.153761 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964316c7-6da5-4725-8140-6d249ee01a68" path="/var/lib/kubelet/pods/964316c7-6da5-4725-8140-6d249ee01a68/volumes" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.246815 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.362641 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-ssh-key\") pod \"701d7b10-8594-480e-8b64-ee2f6513b970\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.362730 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-inventory\") pod \"701d7b10-8594-480e-8b64-ee2f6513b970\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.362761 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5cmh\" (UniqueName: \"kubernetes.io/projected/701d7b10-8594-480e-8b64-ee2f6513b970-kube-api-access-k5cmh\") pod \"701d7b10-8594-480e-8b64-ee2f6513b970\" (UID: \"701d7b10-8594-480e-8b64-ee2f6513b970\") " Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.368196 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/701d7b10-8594-480e-8b64-ee2f6513b970-kube-api-access-k5cmh" (OuterVolumeSpecName: "kube-api-access-k5cmh") pod "701d7b10-8594-480e-8b64-ee2f6513b970" (UID: "701d7b10-8594-480e-8b64-ee2f6513b970"). InnerVolumeSpecName "kube-api-access-k5cmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.390966 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-inventory" (OuterVolumeSpecName: "inventory") pod "701d7b10-8594-480e-8b64-ee2f6513b970" (UID: "701d7b10-8594-480e-8b64-ee2f6513b970"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.396377 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "701d7b10-8594-480e-8b64-ee2f6513b970" (UID: "701d7b10-8594-480e-8b64-ee2f6513b970"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.465161 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.465196 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/701d7b10-8594-480e-8b64-ee2f6513b970-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.465209 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5cmh\" (UniqueName: \"kubernetes.io/projected/701d7b10-8594-480e-8b64-ee2f6513b970-kube-api-access-k5cmh\") on node \"crc\" DevicePath \"\"" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.503017 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" event={"ID":"701d7b10-8594-480e-8b64-ee2f6513b970","Type":"ContainerDied","Data":"1d0789965461bfc41bed515032c7846028a9579b7e33959fb0ddd542a983bf75"} Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.503055 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d0789965461bfc41bed515032c7846028a9579b7e33959fb0ddd542a983bf75" Dec 04 06:40:43 crc kubenswrapper[4685]: I1204 06:40:43.503110 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-97x4w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.030398 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w"] Dec 04 06:40:44 crc kubenswrapper[4685]: E1204 06:40:44.031237 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701d7b10-8594-480e-8b64-ee2f6513b970" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.031282 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="701d7b10-8594-480e-8b64-ee2f6513b970" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.032030 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="701d7b10-8594-480e-8b64-ee2f6513b970" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.033675 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.037082 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.037135 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.037155 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.039048 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.052847 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w"] Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.183179 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.183239 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vthg4\" (UniqueName: \"kubernetes.io/projected/117d2695-776f-4596-9668-d99f7bd3f2bc-kube-api-access-vthg4\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.183643 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.285554 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.285797 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.285826 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vthg4\" (UniqueName: \"kubernetes.io/projected/117d2695-776f-4596-9668-d99f7bd3f2bc-kube-api-access-vthg4\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.291930 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.293823 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.309442 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vthg4\" (UniqueName: \"kubernetes.io/projected/117d2695-776f-4596-9668-d99f7bd3f2bc-kube-api-access-vthg4\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:44 crc kubenswrapper[4685]: I1204 06:40:44.358624 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:40:45 crc kubenswrapper[4685]: I1204 06:40:45.053645 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w"] Dec 04 06:40:45 crc kubenswrapper[4685]: W1204 06:40:45.060006 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod117d2695_776f_4596_9668_d99f7bd3f2bc.slice/crio-9c40821a3d115d79dbe3cc391584bce08b9f74ac26a3c531542b5c1e46044a5e WatchSource:0}: Error finding container 9c40821a3d115d79dbe3cc391584bce08b9f74ac26a3c531542b5c1e46044a5e: Status 404 returned error can't find the container with id 9c40821a3d115d79dbe3cc391584bce08b9f74ac26a3c531542b5c1e46044a5e Dec 04 06:40:45 crc kubenswrapper[4685]: I1204 06:40:45.063681 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:40:45 crc kubenswrapper[4685]: I1204 06:40:45.523709 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" event={"ID":"117d2695-776f-4596-9668-d99f7bd3f2bc","Type":"ContainerStarted","Data":"9c40821a3d115d79dbe3cc391584bce08b9f74ac26a3c531542b5c1e46044a5e"} Dec 04 06:40:46 crc kubenswrapper[4685]: I1204 06:40:46.125972 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:40:46 crc kubenswrapper[4685]: E1204 06:40:46.126648 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:40:46 crc kubenswrapper[4685]: I1204 06:40:46.536212 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" event={"ID":"117d2695-776f-4596-9668-d99f7bd3f2bc","Type":"ContainerStarted","Data":"11f13d01d3b1d7f4c36f888d3a507173864d3b9ec1a3d738f13b0faa04b56bf1"} Dec 04 06:40:46 crc kubenswrapper[4685]: I1204 06:40:46.559653 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" podStartSLOduration=3.370137561 podStartE2EDuration="3.559633618s" podCreationTimestamp="2025-12-04 06:40:43 +0000 UTC" firstStartedPulling="2025-12-04 06:40:45.063429526 +0000 UTC m=+1782.251660301" lastFinishedPulling="2025-12-04 06:40:45.252925573 +0000 UTC m=+1782.441156358" observedRunningTime="2025-12-04 06:40:46.556543022 +0000 UTC m=+1783.744773807" watchObservedRunningTime="2025-12-04 06:40:46.559633618 +0000 UTC m=+1783.747864403" Dec 04 06:40:59 crc kubenswrapper[4685]: I1204 06:40:59.126475 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:40:59 crc kubenswrapper[4685]: E1204 06:40:59.127600 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:41:13 crc kubenswrapper[4685]: I1204 06:41:13.061063 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-nzzq6"] Dec 04 06:41:13 crc kubenswrapper[4685]: I1204 06:41:13.070850 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-nzzq6"] Dec 04 06:41:13 crc kubenswrapper[4685]: I1204 06:41:13.131393 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:41:13 crc kubenswrapper[4685]: E1204 06:41:13.131675 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:41:13 crc kubenswrapper[4685]: I1204 06:41:13.140140 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baf8cf2f-5cdb-45b0-898a-1c7f57330a86" path="/var/lib/kubelet/pods/baf8cf2f-5cdb-45b0-898a-1c7f57330a86/volumes" Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.062057 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-r9p2n"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.086280 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3e55-account-create-update-rtbnz"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.096961 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-sw96z"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.107490 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0f7b-account-create-update-tjxnm"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.116940 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3e55-account-create-update-rtbnz"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.131953 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-sw96z"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.144606 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-r9p2n"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.158711 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-e025-account-create-update-xg92x"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.168976 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0f7b-account-create-update-tjxnm"] Dec 04 06:41:14 crc kubenswrapper[4685]: I1204 06:41:14.176718 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-e025-account-create-update-xg92x"] Dec 04 06:41:15 crc kubenswrapper[4685]: I1204 06:41:15.144523 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21dbe351-9039-4820-a75d-e06bbaba52f4" path="/var/lib/kubelet/pods/21dbe351-9039-4820-a75d-e06bbaba52f4/volumes" Dec 04 06:41:15 crc kubenswrapper[4685]: I1204 06:41:15.146052 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e1d9c5-7b1f-41bb-b58b-190488337724" path="/var/lib/kubelet/pods/26e1d9c5-7b1f-41bb-b58b-190488337724/volumes" Dec 04 06:41:15 crc kubenswrapper[4685]: I1204 06:41:15.147403 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d0095a3-18ba-4bec-b36a-fd9720d76ddd" path="/var/lib/kubelet/pods/2d0095a3-18ba-4bec-b36a-fd9720d76ddd/volumes" Dec 04 06:41:15 crc kubenswrapper[4685]: I1204 06:41:15.149701 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c346e92-8b63-4236-81fc-83864447360a" path="/var/lib/kubelet/pods/4c346e92-8b63-4236-81fc-83864447360a/volumes" Dec 04 06:41:15 crc kubenswrapper[4685]: I1204 06:41:15.152142 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d488956-27a8-4cea-8adf-2bc498bf1868" path="/var/lib/kubelet/pods/4d488956-27a8-4cea-8adf-2bc498bf1868/volumes" Dec 04 06:41:17 crc kubenswrapper[4685]: I1204 06:41:17.820512 4685 scope.go:117] "RemoveContainer" containerID="17385e95af7a11a08f64971163a7c1d699257f4e8d61af5368dbd211eb6c9070" Dec 04 06:41:17 crc kubenswrapper[4685]: I1204 06:41:17.852352 4685 scope.go:117] "RemoveContainer" containerID="f5fc4688f928b73cc2061273f738cf8a8f70956d624c8ba3e567b84ffd763b17" Dec 04 06:41:17 crc kubenswrapper[4685]: I1204 06:41:17.905207 4685 scope.go:117] "RemoveContainer" containerID="d5be38bcad8d1aaf32f17f50810883c8dae3c6fb4a56e660776127539507d9cc" Dec 04 06:41:17 crc kubenswrapper[4685]: I1204 06:41:17.954454 4685 scope.go:117] "RemoveContainer" containerID="66a9b48a97f988823569864e648e76d2467cdf26679169fe46b25d93432385e3" Dec 04 06:41:17 crc kubenswrapper[4685]: I1204 06:41:17.994092 4685 scope.go:117] "RemoveContainer" containerID="dfe3ec87dcc103e750d55dae54a2e0d01f746764027360ddef79b6fa37fbfa75" Dec 04 06:41:18 crc kubenswrapper[4685]: I1204 06:41:18.040107 4685 scope.go:117] "RemoveContainer" containerID="5f45cd1834d62289fa089266d9593c98c8f8ddc608e78e7bca16ae023f56bfb9" Dec 04 06:41:18 crc kubenswrapper[4685]: I1204 06:41:18.072265 4685 scope.go:117] "RemoveContainer" containerID="b8e09f16dce9dcdec259fcc9b33007c3a656dcfd4a71640869312ea4072637dc" Dec 04 06:41:18 crc kubenswrapper[4685]: I1204 06:41:18.102731 4685 scope.go:117] "RemoveContainer" containerID="62ea205bcc848b1fb70edea493bc2bc1106dacc6b3196e4c3d38f7a24555f1de" Dec 04 06:41:18 crc kubenswrapper[4685]: I1204 06:41:18.127872 4685 scope.go:117] "RemoveContainer" containerID="eb76df02727dedb360949a4bd6dcb29986d216ae56c550dc4f2a5c47ab63142d" Dec 04 06:41:18 crc kubenswrapper[4685]: I1204 06:41:18.156421 4685 scope.go:117] "RemoveContainer" containerID="eddd8dea3e1ba7f7c2f7b5188170acda9d987dc8ab31a43c1061e616a86e8b30" Dec 04 06:41:25 crc kubenswrapper[4685]: I1204 06:41:25.126544 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:41:25 crc kubenswrapper[4685]: E1204 06:41:25.127598 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:41:36 crc kubenswrapper[4685]: I1204 06:41:36.126325 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:41:36 crc kubenswrapper[4685]: E1204 06:41:36.127199 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:41:39 crc kubenswrapper[4685]: I1204 06:41:39.041767 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-chrlf"] Dec 04 06:41:39 crc kubenswrapper[4685]: I1204 06:41:39.055360 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-chrlf"] Dec 04 06:41:39 crc kubenswrapper[4685]: I1204 06:41:39.139685 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18ec8793-df06-4f80-9ddc-148283651ef9" path="/var/lib/kubelet/pods/18ec8793-df06-4f80-9ddc-148283651ef9/volumes" Dec 04 06:41:50 crc kubenswrapper[4685]: I1204 06:41:50.126149 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:41:51 crc kubenswrapper[4685]: I1204 06:41:51.265234 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"06d5f8fe304d520b079828f91b96cf934d5f4cad4ffa07e56214997b67c1a8a2"} Dec 04 06:42:01 crc kubenswrapper[4685]: I1204 06:42:01.046845 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-pp2fn"] Dec 04 06:42:01 crc kubenswrapper[4685]: I1204 06:42:01.069983 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-pp2fn"] Dec 04 06:42:01 crc kubenswrapper[4685]: I1204 06:42:01.141460 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8073c264-8fb4-4316-8b24-5de5ca8238be" path="/var/lib/kubelet/pods/8073c264-8fb4-4316-8b24-5de5ca8238be/volumes" Dec 04 06:42:03 crc kubenswrapper[4685]: I1204 06:42:03.037128 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zc8bc"] Dec 04 06:42:03 crc kubenswrapper[4685]: I1204 06:42:03.051635 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zc8bc"] Dec 04 06:42:03 crc kubenswrapper[4685]: I1204 06:42:03.147977 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b37a166d-763c-4492-bcc8-fef7326f7d82" path="/var/lib/kubelet/pods/b37a166d-763c-4492-bcc8-fef7326f7d82/volumes" Dec 04 06:42:03 crc kubenswrapper[4685]: I1204 06:42:03.397442 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" event={"ID":"117d2695-776f-4596-9668-d99f7bd3f2bc","Type":"ContainerDied","Data":"11f13d01d3b1d7f4c36f888d3a507173864d3b9ec1a3d738f13b0faa04b56bf1"} Dec 04 06:42:03 crc kubenswrapper[4685]: I1204 06:42:03.397510 4685 generic.go:334] "Generic (PLEG): container finished" podID="117d2695-776f-4596-9668-d99f7bd3f2bc" containerID="11f13d01d3b1d7f4c36f888d3a507173864d3b9ec1a3d738f13b0faa04b56bf1" exitCode=0 Dec 04 06:42:04 crc kubenswrapper[4685]: I1204 06:42:04.969196 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.119530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-ssh-key\") pod \"117d2695-776f-4596-9668-d99f7bd3f2bc\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.119767 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vthg4\" (UniqueName: \"kubernetes.io/projected/117d2695-776f-4596-9668-d99f7bd3f2bc-kube-api-access-vthg4\") pod \"117d2695-776f-4596-9668-d99f7bd3f2bc\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.119809 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-inventory\") pod \"117d2695-776f-4596-9668-d99f7bd3f2bc\" (UID: \"117d2695-776f-4596-9668-d99f7bd3f2bc\") " Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.125279 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/117d2695-776f-4596-9668-d99f7bd3f2bc-kube-api-access-vthg4" (OuterVolumeSpecName: "kube-api-access-vthg4") pod "117d2695-776f-4596-9668-d99f7bd3f2bc" (UID: "117d2695-776f-4596-9668-d99f7bd3f2bc"). InnerVolumeSpecName "kube-api-access-vthg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.171567 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "117d2695-776f-4596-9668-d99f7bd3f2bc" (UID: "117d2695-776f-4596-9668-d99f7bd3f2bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.175605 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-inventory" (OuterVolumeSpecName: "inventory") pod "117d2695-776f-4596-9668-d99f7bd3f2bc" (UID: "117d2695-776f-4596-9668-d99f7bd3f2bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.223000 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vthg4\" (UniqueName: \"kubernetes.io/projected/117d2695-776f-4596-9668-d99f7bd3f2bc-kube-api-access-vthg4\") on node \"crc\" DevicePath \"\"" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.223058 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.223078 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/117d2695-776f-4596-9668-d99f7bd3f2bc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.424022 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" event={"ID":"117d2695-776f-4596-9668-d99f7bd3f2bc","Type":"ContainerDied","Data":"9c40821a3d115d79dbe3cc391584bce08b9f74ac26a3c531542b5c1e46044a5e"} Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.424064 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c40821a3d115d79dbe3cc391584bce08b9f74ac26a3c531542b5c1e46044a5e" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.424124 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.531198 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd"] Dec 04 06:42:05 crc kubenswrapper[4685]: E1204 06:42:05.531715 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="117d2695-776f-4596-9668-d99f7bd3f2bc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.531738 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="117d2695-776f-4596-9668-d99f7bd3f2bc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.531983 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="117d2695-776f-4596-9668-d99f7bd3f2bc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.532845 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.536102 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.536947 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.537356 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.541800 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.547155 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd"] Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.631216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.631710 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh2p5\" (UniqueName: \"kubernetes.io/projected/080fc1a1-8d27-4b03-bd8f-04d200f62da2-kube-api-access-fh2p5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.631793 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.734001 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh2p5\" (UniqueName: \"kubernetes.io/projected/080fc1a1-8d27-4b03-bd8f-04d200f62da2-kube-api-access-fh2p5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.734095 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.734167 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.739322 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.745469 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.755576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh2p5\" (UniqueName: \"kubernetes.io/projected/080fc1a1-8d27-4b03-bd8f-04d200f62da2-kube-api-access-fh2p5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:05 crc kubenswrapper[4685]: I1204 06:42:05.851781 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:06 crc kubenswrapper[4685]: I1204 06:42:06.237302 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd"] Dec 04 06:42:06 crc kubenswrapper[4685]: I1204 06:42:06.433557 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" event={"ID":"080fc1a1-8d27-4b03-bd8f-04d200f62da2","Type":"ContainerStarted","Data":"7ac8bb2b8a4607163ef0c15c3595eb3dca477979c04b7aa62522b0de2e93e625"} Dec 04 06:42:07 crc kubenswrapper[4685]: I1204 06:42:07.444359 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" event={"ID":"080fc1a1-8d27-4b03-bd8f-04d200f62da2","Type":"ContainerStarted","Data":"4ee2f99069c6119b49079588d849a43f59764b1b35d4cf0eb657b8fe857b0bfe"} Dec 04 06:42:12 crc kubenswrapper[4685]: I1204 06:42:12.500091 4685 generic.go:334] "Generic (PLEG): container finished" podID="080fc1a1-8d27-4b03-bd8f-04d200f62da2" containerID="4ee2f99069c6119b49079588d849a43f59764b1b35d4cf0eb657b8fe857b0bfe" exitCode=0 Dec 04 06:42:12 crc kubenswrapper[4685]: I1204 06:42:12.500184 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" event={"ID":"080fc1a1-8d27-4b03-bd8f-04d200f62da2","Type":"ContainerDied","Data":"4ee2f99069c6119b49079588d849a43f59764b1b35d4cf0eb657b8fe857b0bfe"} Dec 04 06:42:13 crc kubenswrapper[4685]: I1204 06:42:13.941302 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.112328 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh2p5\" (UniqueName: \"kubernetes.io/projected/080fc1a1-8d27-4b03-bd8f-04d200f62da2-kube-api-access-fh2p5\") pod \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.112460 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-ssh-key\") pod \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.112621 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-inventory\") pod \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\" (UID: \"080fc1a1-8d27-4b03-bd8f-04d200f62da2\") " Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.117940 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/080fc1a1-8d27-4b03-bd8f-04d200f62da2-kube-api-access-fh2p5" (OuterVolumeSpecName: "kube-api-access-fh2p5") pod "080fc1a1-8d27-4b03-bd8f-04d200f62da2" (UID: "080fc1a1-8d27-4b03-bd8f-04d200f62da2"). InnerVolumeSpecName "kube-api-access-fh2p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.139221 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-inventory" (OuterVolumeSpecName: "inventory") pod "080fc1a1-8d27-4b03-bd8f-04d200f62da2" (UID: "080fc1a1-8d27-4b03-bd8f-04d200f62da2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.147524 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "080fc1a1-8d27-4b03-bd8f-04d200f62da2" (UID: "080fc1a1-8d27-4b03-bd8f-04d200f62da2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.215314 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.215393 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh2p5\" (UniqueName: \"kubernetes.io/projected/080fc1a1-8d27-4b03-bd8f-04d200f62da2-kube-api-access-fh2p5\") on node \"crc\" DevicePath \"\"" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.215457 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/080fc1a1-8d27-4b03-bd8f-04d200f62da2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.517579 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" event={"ID":"080fc1a1-8d27-4b03-bd8f-04d200f62da2","Type":"ContainerDied","Data":"7ac8bb2b8a4607163ef0c15c3595eb3dca477979c04b7aa62522b0de2e93e625"} Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.517843 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ac8bb2b8a4607163ef0c15c3595eb3dca477979c04b7aa62522b0de2e93e625" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.517656 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.581507 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq"] Dec 04 06:42:14 crc kubenswrapper[4685]: E1204 06:42:14.582086 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080fc1a1-8d27-4b03-bd8f-04d200f62da2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.582148 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="080fc1a1-8d27-4b03-bd8f-04d200f62da2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.582393 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="080fc1a1-8d27-4b03-bd8f-04d200f62da2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.583098 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.585093 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.585556 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.585740 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.585896 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.590779 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq"] Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.724687 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njpk2\" (UniqueName: \"kubernetes.io/projected/d64ce345-0870-44f8-8cfb-da7b8cd6da34-kube-api-access-njpk2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.724957 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.725094 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.826838 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njpk2\" (UniqueName: \"kubernetes.io/projected/d64ce345-0870-44f8-8cfb-da7b8cd6da34-kube-api-access-njpk2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.827210 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.827476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.831176 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.837819 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.846219 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njpk2\" (UniqueName: \"kubernetes.io/projected/d64ce345-0870-44f8-8cfb-da7b8cd6da34-kube-api-access-njpk2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r22kq\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:14 crc kubenswrapper[4685]: I1204 06:42:14.900109 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:42:15 crc kubenswrapper[4685]: I1204 06:42:15.485304 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq"] Dec 04 06:42:15 crc kubenswrapper[4685]: I1204 06:42:15.528701 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" event={"ID":"d64ce345-0870-44f8-8cfb-da7b8cd6da34","Type":"ContainerStarted","Data":"2f1fe430c7f36103a6bd8844637bbd8b9b574d01f6ddae823a987807bbb0b422"} Dec 04 06:42:16 crc kubenswrapper[4685]: I1204 06:42:16.554744 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" event={"ID":"d64ce345-0870-44f8-8cfb-da7b8cd6da34","Type":"ContainerStarted","Data":"ce53acaa76819d88c0970080878b36c9ce508db7a3b4d7a2eb586d82549d3a34"} Dec 04 06:42:16 crc kubenswrapper[4685]: I1204 06:42:16.582614 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" podStartSLOduration=2.442466921 podStartE2EDuration="2.582579689s" podCreationTimestamp="2025-12-04 06:42:14 +0000 UTC" firstStartedPulling="2025-12-04 06:42:15.497506413 +0000 UTC m=+1872.685737208" lastFinishedPulling="2025-12-04 06:42:15.637619201 +0000 UTC m=+1872.825849976" observedRunningTime="2025-12-04 06:42:16.568936334 +0000 UTC m=+1873.757167149" watchObservedRunningTime="2025-12-04 06:42:16.582579689 +0000 UTC m=+1873.770810544" Dec 04 06:42:18 crc kubenswrapper[4685]: I1204 06:42:18.384364 4685 scope.go:117] "RemoveContainer" containerID="d3789629a7d33764b1db5b73e19ab06c1ea82860b04a1c59b836bba229fcfaf9" Dec 04 06:42:18 crc kubenswrapper[4685]: I1204 06:42:18.426993 4685 scope.go:117] "RemoveContainer" containerID="cb57c41e0d01aacf3bce298f5de0330ebb3d9388841c18402625385e47f3efd8" Dec 04 06:42:18 crc kubenswrapper[4685]: I1204 06:42:18.513601 4685 scope.go:117] "RemoveContainer" containerID="b857ed3a1b6e515b527451457d9dd77d94d20b4ced96da049ad39e13b342e285" Dec 04 06:42:45 crc kubenswrapper[4685]: I1204 06:42:45.053279 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-jzvtg"] Dec 04 06:42:45 crc kubenswrapper[4685]: I1204 06:42:45.066726 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-jzvtg"] Dec 04 06:42:45 crc kubenswrapper[4685]: I1204 06:42:45.138098 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131efcaa-5126-43cc-afa2-264f6c8bee70" path="/var/lib/kubelet/pods/131efcaa-5126-43cc-afa2-264f6c8bee70/volumes" Dec 04 06:43:00 crc kubenswrapper[4685]: I1204 06:42:59.999378 4685 generic.go:334] "Generic (PLEG): container finished" podID="d64ce345-0870-44f8-8cfb-da7b8cd6da34" containerID="ce53acaa76819d88c0970080878b36c9ce508db7a3b4d7a2eb586d82549d3a34" exitCode=0 Dec 04 06:43:00 crc kubenswrapper[4685]: I1204 06:42:59.999482 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" event={"ID":"d64ce345-0870-44f8-8cfb-da7b8cd6da34","Type":"ContainerDied","Data":"ce53acaa76819d88c0970080878b36c9ce508db7a3b4d7a2eb586d82549d3a34"} Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.521032 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.540398 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njpk2\" (UniqueName: \"kubernetes.io/projected/d64ce345-0870-44f8-8cfb-da7b8cd6da34-kube-api-access-njpk2\") pod \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.540574 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-inventory\") pod \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.540697 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-ssh-key\") pod \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\" (UID: \"d64ce345-0870-44f8-8cfb-da7b8cd6da34\") " Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.549885 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64ce345-0870-44f8-8cfb-da7b8cd6da34-kube-api-access-njpk2" (OuterVolumeSpecName: "kube-api-access-njpk2") pod "d64ce345-0870-44f8-8cfb-da7b8cd6da34" (UID: "d64ce345-0870-44f8-8cfb-da7b8cd6da34"). InnerVolumeSpecName "kube-api-access-njpk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.572178 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d64ce345-0870-44f8-8cfb-da7b8cd6da34" (UID: "d64ce345-0870-44f8-8cfb-da7b8cd6da34"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.588906 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-inventory" (OuterVolumeSpecName: "inventory") pod "d64ce345-0870-44f8-8cfb-da7b8cd6da34" (UID: "d64ce345-0870-44f8-8cfb-da7b8cd6da34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.642121 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njpk2\" (UniqueName: \"kubernetes.io/projected/d64ce345-0870-44f8-8cfb-da7b8cd6da34-kube-api-access-njpk2\") on node \"crc\" DevicePath \"\"" Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.642218 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:43:01 crc kubenswrapper[4685]: I1204 06:43:01.642229 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64ce345-0870-44f8-8cfb-da7b8cd6da34-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.028853 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" event={"ID":"d64ce345-0870-44f8-8cfb-da7b8cd6da34","Type":"ContainerDied","Data":"2f1fe430c7f36103a6bd8844637bbd8b9b574d01f6ddae823a987807bbb0b422"} Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.028905 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f1fe430c7f36103a6bd8844637bbd8b9b574d01f6ddae823a987807bbb0b422" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.028944 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r22kq" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.116908 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng"] Dec 04 06:43:02 crc kubenswrapper[4685]: E1204 06:43:02.117289 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ce345-0870-44f8-8cfb-da7b8cd6da34" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.117304 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ce345-0870-44f8-8cfb-da7b8cd6da34" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.117527 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64ce345-0870-44f8-8cfb-da7b8cd6da34" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.118278 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.122873 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.122967 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.123023 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.122985 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.147674 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng"] Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.154060 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.154298 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.154729 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgm98\" (UniqueName: \"kubernetes.io/projected/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-kube-api-access-wgm98\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.256609 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.256725 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.256780 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgm98\" (UniqueName: \"kubernetes.io/projected/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-kube-api-access-wgm98\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.262273 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.266930 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.273205 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgm98\" (UniqueName: \"kubernetes.io/projected/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-kube-api-access-wgm98\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:02 crc kubenswrapper[4685]: I1204 06:43:02.438396 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:43:03 crc kubenswrapper[4685]: I1204 06:43:03.052550 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng"] Dec 04 06:43:03 crc kubenswrapper[4685]: I1204 06:43:03.247245 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:43:04 crc kubenswrapper[4685]: I1204 06:43:04.047992 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" event={"ID":"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05","Type":"ContainerStarted","Data":"b065b593ea79cbb11d0e3390bc5edb65ee610f8a061f82dc7ad858079a56b38b"} Dec 04 06:43:04 crc kubenswrapper[4685]: I1204 06:43:04.049066 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" event={"ID":"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05","Type":"ContainerStarted","Data":"2a47eb203584afd1a55dc688e43bedb0cf41615b7ec51d0fc741129e63a82d3b"} Dec 04 06:43:04 crc kubenswrapper[4685]: I1204 06:43:04.080722 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" podStartSLOduration=1.8958030259999998 podStartE2EDuration="2.080695715s" podCreationTimestamp="2025-12-04 06:43:02 +0000 UTC" firstStartedPulling="2025-12-04 06:43:03.059204421 +0000 UTC m=+1920.247435186" lastFinishedPulling="2025-12-04 06:43:03.24409709 +0000 UTC m=+1920.432327875" observedRunningTime="2025-12-04 06:43:04.069528525 +0000 UTC m=+1921.257759330" watchObservedRunningTime="2025-12-04 06:43:04.080695715 +0000 UTC m=+1921.268926490" Dec 04 06:43:18 crc kubenswrapper[4685]: I1204 06:43:18.602815 4685 scope.go:117] "RemoveContainer" containerID="a2d5a66fde3064601e12843baf9f2145e855b721a21ce447bf3769ad352fc0af" Dec 04 06:44:02 crc kubenswrapper[4685]: I1204 06:44:02.660574 4685 generic.go:334] "Generic (PLEG): container finished" podID="2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" containerID="b065b593ea79cbb11d0e3390bc5edb65ee610f8a061f82dc7ad858079a56b38b" exitCode=0 Dec 04 06:44:02 crc kubenswrapper[4685]: I1204 06:44:02.660717 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" event={"ID":"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05","Type":"ContainerDied","Data":"b065b593ea79cbb11d0e3390bc5edb65ee610f8a061f82dc7ad858079a56b38b"} Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.069631 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.101532 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgm98\" (UniqueName: \"kubernetes.io/projected/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-kube-api-access-wgm98\") pod \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.101702 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-ssh-key\") pod \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.101782 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-inventory\") pod \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\" (UID: \"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05\") " Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.109315 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-kube-api-access-wgm98" (OuterVolumeSpecName: "kube-api-access-wgm98") pod "2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" (UID: "2729f4b1-7245-40d2-a0e0-f4bb72e2fb05"). InnerVolumeSpecName "kube-api-access-wgm98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.130370 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" (UID: "2729f4b1-7245-40d2-a0e0-f4bb72e2fb05"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.135231 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-inventory" (OuterVolumeSpecName: "inventory") pod "2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" (UID: "2729f4b1-7245-40d2-a0e0-f4bb72e2fb05"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.203627 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgm98\" (UniqueName: \"kubernetes.io/projected/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-kube-api-access-wgm98\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.203679 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.203696 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2729f4b1-7245-40d2-a0e0-f4bb72e2fb05-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.693618 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" event={"ID":"2729f4b1-7245-40d2-a0e0-f4bb72e2fb05","Type":"ContainerDied","Data":"2a47eb203584afd1a55dc688e43bedb0cf41615b7ec51d0fc741129e63a82d3b"} Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.693963 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a47eb203584afd1a55dc688e43bedb0cf41615b7ec51d0fc741129e63a82d3b" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.693709 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.815925 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xldx9"] Dec 04 06:44:04 crc kubenswrapper[4685]: E1204 06:44:04.817009 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.817107 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.817450 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2729f4b1-7245-40d2-a0e0-f4bb72e2fb05" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.818364 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.823200 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.823635 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.823672 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.824751 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xldx9"] Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.824977 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.916221 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccb7m\" (UniqueName: \"kubernetes.io/projected/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-kube-api-access-ccb7m\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.916394 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:04 crc kubenswrapper[4685]: I1204 06:44:04.916711 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.018787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.018945 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccb7m\" (UniqueName: \"kubernetes.io/projected/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-kube-api-access-ccb7m\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.019144 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.024896 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.025075 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.042219 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccb7m\" (UniqueName: \"kubernetes.io/projected/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-kube-api-access-ccb7m\") pod \"ssh-known-hosts-edpm-deployment-xldx9\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.133765 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:05 crc kubenswrapper[4685]: I1204 06:44:05.819067 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xldx9"] Dec 04 06:44:06 crc kubenswrapper[4685]: I1204 06:44:06.120760 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:44:06 crc kubenswrapper[4685]: I1204 06:44:06.121078 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:44:06 crc kubenswrapper[4685]: I1204 06:44:06.727495 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" event={"ID":"f33640dd-287a-4de5-be7a-bc4fa6fd72fe","Type":"ContainerStarted","Data":"c68dd6960b0c85c0112f653352f49bc8d15c1f8ce24ec2b5d1e9203a55c74534"} Dec 04 06:44:06 crc kubenswrapper[4685]: I1204 06:44:06.727580 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" event={"ID":"f33640dd-287a-4de5-be7a-bc4fa6fd72fe","Type":"ContainerStarted","Data":"607247e8c3ed18f4d4445cbeb9141812ef29c213f4a6a306c307b500206fbd73"} Dec 04 06:44:06 crc kubenswrapper[4685]: I1204 06:44:06.756367 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" podStartSLOduration=2.563066044 podStartE2EDuration="2.756348605s" podCreationTimestamp="2025-12-04 06:44:04 +0000 UTC" firstStartedPulling="2025-12-04 06:44:05.819942799 +0000 UTC m=+1983.008173584" lastFinishedPulling="2025-12-04 06:44:06.01322537 +0000 UTC m=+1983.201456145" observedRunningTime="2025-12-04 06:44:06.749224091 +0000 UTC m=+1983.937454886" watchObservedRunningTime="2025-12-04 06:44:06.756348605 +0000 UTC m=+1983.944579380" Dec 04 06:44:13 crc kubenswrapper[4685]: E1204 06:44:13.672404 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf33640dd_287a_4de5_be7a_bc4fa6fd72fe.slice/crio-c68dd6960b0c85c0112f653352f49bc8d15c1f8ce24ec2b5d1e9203a55c74534.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf33640dd_287a_4de5_be7a_bc4fa6fd72fe.slice/crio-conmon-c68dd6960b0c85c0112f653352f49bc8d15c1f8ce24ec2b5d1e9203a55c74534.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:44:13 crc kubenswrapper[4685]: I1204 06:44:13.805115 4685 generic.go:334] "Generic (PLEG): container finished" podID="f33640dd-287a-4de5-be7a-bc4fa6fd72fe" containerID="c68dd6960b0c85c0112f653352f49bc8d15c1f8ce24ec2b5d1e9203a55c74534" exitCode=0 Dec 04 06:44:13 crc kubenswrapper[4685]: I1204 06:44:13.805259 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" event={"ID":"f33640dd-287a-4de5-be7a-bc4fa6fd72fe","Type":"ContainerDied","Data":"c68dd6960b0c85c0112f653352f49bc8d15c1f8ce24ec2b5d1e9203a55c74534"} Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.322282 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.417191 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-ssh-key-openstack-edpm-ipam\") pod \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.417293 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccb7m\" (UniqueName: \"kubernetes.io/projected/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-kube-api-access-ccb7m\") pod \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.417336 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-inventory-0\") pod \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\" (UID: \"f33640dd-287a-4de5-be7a-bc4fa6fd72fe\") " Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.425947 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-kube-api-access-ccb7m" (OuterVolumeSpecName: "kube-api-access-ccb7m") pod "f33640dd-287a-4de5-be7a-bc4fa6fd72fe" (UID: "f33640dd-287a-4de5-be7a-bc4fa6fd72fe"). InnerVolumeSpecName "kube-api-access-ccb7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.446593 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "f33640dd-287a-4de5-be7a-bc4fa6fd72fe" (UID: "f33640dd-287a-4de5-be7a-bc4fa6fd72fe"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.458156 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f33640dd-287a-4de5-be7a-bc4fa6fd72fe" (UID: "f33640dd-287a-4de5-be7a-bc4fa6fd72fe"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.519936 4685 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.520263 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.520277 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccb7m\" (UniqueName: \"kubernetes.io/projected/f33640dd-287a-4de5-be7a-bc4fa6fd72fe-kube-api-access-ccb7m\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.822869 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" event={"ID":"f33640dd-287a-4de5-be7a-bc4fa6fd72fe","Type":"ContainerDied","Data":"607247e8c3ed18f4d4445cbeb9141812ef29c213f4a6a306c307b500206fbd73"} Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.822908 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xldx9" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.822916 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="607247e8c3ed18f4d4445cbeb9141812ef29c213f4a6a306c307b500206fbd73" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.929769 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk"] Dec 04 06:44:15 crc kubenswrapper[4685]: E1204 06:44:15.930315 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33640dd-287a-4de5-be7a-bc4fa6fd72fe" containerName="ssh-known-hosts-edpm-deployment" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.930337 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33640dd-287a-4de5-be7a-bc4fa6fd72fe" containerName="ssh-known-hosts-edpm-deployment" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.930559 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33640dd-287a-4de5-be7a-bc4fa6fd72fe" containerName="ssh-known-hosts-edpm-deployment" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.931356 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.936127 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.936123 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.936591 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.936644 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:44:15 crc kubenswrapper[4685]: I1204 06:44:15.947070 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk"] Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.028233 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.028490 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.028738 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbgv8\" (UniqueName: \"kubernetes.io/projected/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-kube-api-access-cbgv8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.130489 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.130606 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbgv8\" (UniqueName: \"kubernetes.io/projected/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-kube-api-access-cbgv8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.130660 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.135322 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.143671 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.154226 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbgv8\" (UniqueName: \"kubernetes.io/projected/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-kube-api-access-cbgv8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rxvkk\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.265568 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:16 crc kubenswrapper[4685]: I1204 06:44:16.825862 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk"] Dec 04 06:44:17 crc kubenswrapper[4685]: I1204 06:44:17.841649 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" event={"ID":"d20c1b8a-7db1-4d00-8907-88a6a3e3d312","Type":"ContainerStarted","Data":"edcbde3cd48c0078c316c537f40d602739f67a813588a10bf18a5b3a0e4294c8"} Dec 04 06:44:17 crc kubenswrapper[4685]: I1204 06:44:17.842219 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" event={"ID":"d20c1b8a-7db1-4d00-8907-88a6a3e3d312","Type":"ContainerStarted","Data":"6f85711edb08bde34fe80960c68082c422c7ba67359548799b69f6d2eb6a1eda"} Dec 04 06:44:17 crc kubenswrapper[4685]: I1204 06:44:17.867587 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" podStartSLOduration=2.6900241620000003 podStartE2EDuration="2.86756441s" podCreationTimestamp="2025-12-04 06:44:15 +0000 UTC" firstStartedPulling="2025-12-04 06:44:16.833786921 +0000 UTC m=+1994.022017696" lastFinishedPulling="2025-12-04 06:44:17.011327169 +0000 UTC m=+1994.199557944" observedRunningTime="2025-12-04 06:44:17.862025477 +0000 UTC m=+1995.050256282" watchObservedRunningTime="2025-12-04 06:44:17.86756441 +0000 UTC m=+1995.055795195" Dec 04 06:44:25 crc kubenswrapper[4685]: I1204 06:44:25.931004 4685 generic.go:334] "Generic (PLEG): container finished" podID="d20c1b8a-7db1-4d00-8907-88a6a3e3d312" containerID="edcbde3cd48c0078c316c537f40d602739f67a813588a10bf18a5b3a0e4294c8" exitCode=0 Dec 04 06:44:25 crc kubenswrapper[4685]: I1204 06:44:25.931149 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" event={"ID":"d20c1b8a-7db1-4d00-8907-88a6a3e3d312","Type":"ContainerDied","Data":"edcbde3cd48c0078c316c537f40d602739f67a813588a10bf18a5b3a0e4294c8"} Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.330896 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.469995 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-inventory\") pod \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.470152 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-ssh-key\") pod \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.470208 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbgv8\" (UniqueName: \"kubernetes.io/projected/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-kube-api-access-cbgv8\") pod \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\" (UID: \"d20c1b8a-7db1-4d00-8907-88a6a3e3d312\") " Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.478027 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-kube-api-access-cbgv8" (OuterVolumeSpecName: "kube-api-access-cbgv8") pod "d20c1b8a-7db1-4d00-8907-88a6a3e3d312" (UID: "d20c1b8a-7db1-4d00-8907-88a6a3e3d312"). InnerVolumeSpecName "kube-api-access-cbgv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.500937 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-inventory" (OuterVolumeSpecName: "inventory") pod "d20c1b8a-7db1-4d00-8907-88a6a3e3d312" (UID: "d20c1b8a-7db1-4d00-8907-88a6a3e3d312"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.520882 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d20c1b8a-7db1-4d00-8907-88a6a3e3d312" (UID: "d20c1b8a-7db1-4d00-8907-88a6a3e3d312"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.573324 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbgv8\" (UniqueName: \"kubernetes.io/projected/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-kube-api-access-cbgv8\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.573370 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.573383 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d20c1b8a-7db1-4d00-8907-88a6a3e3d312-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.970166 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" event={"ID":"d20c1b8a-7db1-4d00-8907-88a6a3e3d312","Type":"ContainerDied","Data":"6f85711edb08bde34fe80960c68082c422c7ba67359548799b69f6d2eb6a1eda"} Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.970708 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f85711edb08bde34fe80960c68082c422c7ba67359548799b69f6d2eb6a1eda" Dec 04 06:44:27 crc kubenswrapper[4685]: I1204 06:44:27.970278 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rxvkk" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.069457 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp"] Dec 04 06:44:28 crc kubenswrapper[4685]: E1204 06:44:28.069952 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20c1b8a-7db1-4d00-8907-88a6a3e3d312" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.069968 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20c1b8a-7db1-4d00-8907-88a6a3e3d312" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.070215 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20c1b8a-7db1-4d00-8907-88a6a3e3d312" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.070993 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.075637 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.075815 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.075849 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.075927 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.083528 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp"] Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.085128 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2z8s\" (UniqueName: \"kubernetes.io/projected/f64013bb-a12d-4cfd-93de-741c83048e77-kube-api-access-c2z8s\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.085198 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.085357 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.186435 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.186600 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2z8s\" (UniqueName: \"kubernetes.io/projected/f64013bb-a12d-4cfd-93de-741c83048e77-kube-api-access-c2z8s\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.186632 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.190747 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.191142 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.204974 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2z8s\" (UniqueName: \"kubernetes.io/projected/f64013bb-a12d-4cfd-93de-741c83048e77-kube-api-access-c2z8s\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:28 crc kubenswrapper[4685]: I1204 06:44:28.408889 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:29 crc kubenswrapper[4685]: I1204 06:44:29.025817 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp"] Dec 04 06:44:29 crc kubenswrapper[4685]: I1204 06:44:29.990397 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" event={"ID":"f64013bb-a12d-4cfd-93de-741c83048e77","Type":"ContainerStarted","Data":"96cbebb54018e5ef5c621f404526c3f3057df7541c08151b3605cb09becd806b"} Dec 04 06:44:29 crc kubenswrapper[4685]: I1204 06:44:29.990744 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" event={"ID":"f64013bb-a12d-4cfd-93de-741c83048e77","Type":"ContainerStarted","Data":"a7ecc5cf8979926c67762e6be9a025e6247560db1dc460191e9a5b20ba4df0cf"} Dec 04 06:44:30 crc kubenswrapper[4685]: I1204 06:44:30.008697 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" podStartSLOduration=1.8300039510000001 podStartE2EDuration="2.008681984s" podCreationTimestamp="2025-12-04 06:44:28 +0000 UTC" firstStartedPulling="2025-12-04 06:44:29.007373303 +0000 UTC m=+2006.195604088" lastFinishedPulling="2025-12-04 06:44:29.186051336 +0000 UTC m=+2006.374282121" observedRunningTime="2025-12-04 06:44:30.00506857 +0000 UTC m=+2007.193299345" watchObservedRunningTime="2025-12-04 06:44:30.008681984 +0000 UTC m=+2007.196912759" Dec 04 06:44:36 crc kubenswrapper[4685]: I1204 06:44:36.121384 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:44:36 crc kubenswrapper[4685]: I1204 06:44:36.122219 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:44:40 crc kubenswrapper[4685]: I1204 06:44:40.092807 4685 generic.go:334] "Generic (PLEG): container finished" podID="f64013bb-a12d-4cfd-93de-741c83048e77" containerID="96cbebb54018e5ef5c621f404526c3f3057df7541c08151b3605cb09becd806b" exitCode=0 Dec 04 06:44:40 crc kubenswrapper[4685]: I1204 06:44:40.092923 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" event={"ID":"f64013bb-a12d-4cfd-93de-741c83048e77","Type":"ContainerDied","Data":"96cbebb54018e5ef5c621f404526c3f3057df7541c08151b3605cb09becd806b"} Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.574174 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.752323 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-ssh-key\") pod \"f64013bb-a12d-4cfd-93de-741c83048e77\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.752401 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2z8s\" (UniqueName: \"kubernetes.io/projected/f64013bb-a12d-4cfd-93de-741c83048e77-kube-api-access-c2z8s\") pod \"f64013bb-a12d-4cfd-93de-741c83048e77\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.752592 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-inventory\") pod \"f64013bb-a12d-4cfd-93de-741c83048e77\" (UID: \"f64013bb-a12d-4cfd-93de-741c83048e77\") " Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.760771 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f64013bb-a12d-4cfd-93de-741c83048e77-kube-api-access-c2z8s" (OuterVolumeSpecName: "kube-api-access-c2z8s") pod "f64013bb-a12d-4cfd-93de-741c83048e77" (UID: "f64013bb-a12d-4cfd-93de-741c83048e77"). InnerVolumeSpecName "kube-api-access-c2z8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.783121 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f64013bb-a12d-4cfd-93de-741c83048e77" (UID: "f64013bb-a12d-4cfd-93de-741c83048e77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.783273 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-inventory" (OuterVolumeSpecName: "inventory") pod "f64013bb-a12d-4cfd-93de-741c83048e77" (UID: "f64013bb-a12d-4cfd-93de-741c83048e77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.855823 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.855879 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f64013bb-a12d-4cfd-93de-741c83048e77-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:41 crc kubenswrapper[4685]: I1204 06:44:41.855897 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2z8s\" (UniqueName: \"kubernetes.io/projected/f64013bb-a12d-4cfd-93de-741c83048e77-kube-api-access-c2z8s\") on node \"crc\" DevicePath \"\"" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.113016 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" event={"ID":"f64013bb-a12d-4cfd-93de-741c83048e77","Type":"ContainerDied","Data":"a7ecc5cf8979926c67762e6be9a025e6247560db1dc460191e9a5b20ba4df0cf"} Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.113086 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ecc5cf8979926c67762e6be9a025e6247560db1dc460191e9a5b20ba4df0cf" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.113089 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.219927 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq"] Dec 04 06:44:42 crc kubenswrapper[4685]: E1204 06:44:42.220454 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f64013bb-a12d-4cfd-93de-741c83048e77" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.220470 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f64013bb-a12d-4cfd-93de-741c83048e77" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.220737 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f64013bb-a12d-4cfd-93de-741c83048e77" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.221620 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.224750 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.224971 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.225225 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.225396 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.225628 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.225820 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.225948 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.226105 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.229218 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq"] Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.365932 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.365970 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.365991 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366032 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366064 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366126 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366215 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366399 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366548 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366617 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366707 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366767 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bscxq\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-kube-api-access-bscxq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366800 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.366868 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469195 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469264 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469479 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469540 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469594 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469637 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bscxq\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-kube-api-access-bscxq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469678 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469720 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469779 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469841 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469875 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469932 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.469976 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.476388 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.477237 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.476392 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.476542 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.477203 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.477137 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.478442 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.478495 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.478742 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.481093 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.490104 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.490393 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.494269 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.495106 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bscxq\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-kube-api-access-bscxq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:42 crc kubenswrapper[4685]: I1204 06:44:42.553562 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:44:43 crc kubenswrapper[4685]: I1204 06:44:43.148077 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq"] Dec 04 06:44:44 crc kubenswrapper[4685]: I1204 06:44:44.142695 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" event={"ID":"e75fdf41-c07d-4fdd-9abc-334e2164b96e","Type":"ContainerStarted","Data":"05fd69458eb3a713ec3aa4f8113fba05b5d9657289bec1bd9925b50803b4348b"} Dec 04 06:44:44 crc kubenswrapper[4685]: I1204 06:44:44.143720 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" event={"ID":"e75fdf41-c07d-4fdd-9abc-334e2164b96e","Type":"ContainerStarted","Data":"7199c3c962ea0906458b439c52b5e44e5c806e425eb1f5a657913db7504bf9f8"} Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.156499 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" podStartSLOduration=17.956566093 podStartE2EDuration="18.156477303s" podCreationTimestamp="2025-12-04 06:44:42 +0000 UTC" firstStartedPulling="2025-12-04 06:44:43.166580603 +0000 UTC m=+2020.354811398" lastFinishedPulling="2025-12-04 06:44:43.366491793 +0000 UTC m=+2020.554722608" observedRunningTime="2025-12-04 06:44:44.184483854 +0000 UTC m=+2021.372714669" watchObservedRunningTime="2025-12-04 06:45:00.156477303 +0000 UTC m=+2037.344708088" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.162591 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz"] Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.165383 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.168562 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.168837 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.177048 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz"] Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.265650 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f322dc6e-91ac-4e74-995e-53ce399f8b82-config-volume\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.266008 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f322dc6e-91ac-4e74-995e-53ce399f8b82-secret-volume\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.266117 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9226\" (UniqueName: \"kubernetes.io/projected/f322dc6e-91ac-4e74-995e-53ce399f8b82-kube-api-access-m9226\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.368325 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f322dc6e-91ac-4e74-995e-53ce399f8b82-config-volume\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.368381 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f322dc6e-91ac-4e74-995e-53ce399f8b82-secret-volume\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.368405 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9226\" (UniqueName: \"kubernetes.io/projected/f322dc6e-91ac-4e74-995e-53ce399f8b82-kube-api-access-m9226\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.369838 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f322dc6e-91ac-4e74-995e-53ce399f8b82-config-volume\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.377915 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f322dc6e-91ac-4e74-995e-53ce399f8b82-secret-volume\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.384912 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9226\" (UniqueName: \"kubernetes.io/projected/f322dc6e-91ac-4e74-995e-53ce399f8b82-kube-api-access-m9226\") pod \"collect-profiles-29413845-7hmwz\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:00 crc kubenswrapper[4685]: I1204 06:45:00.525893 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:01 crc kubenswrapper[4685]: W1204 06:45:01.021925 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf322dc6e_91ac_4e74_995e_53ce399f8b82.slice/crio-faac2b560e6e3164af3ba5a35b3d3b80c3c47759eb055c3570653df809e520ee WatchSource:0}: Error finding container faac2b560e6e3164af3ba5a35b3d3b80c3c47759eb055c3570653df809e520ee: Status 404 returned error can't find the container with id faac2b560e6e3164af3ba5a35b3d3b80c3c47759eb055c3570653df809e520ee Dec 04 06:45:01 crc kubenswrapper[4685]: I1204 06:45:01.022026 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz"] Dec 04 06:45:01 crc kubenswrapper[4685]: I1204 06:45:01.320072 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" event={"ID":"f322dc6e-91ac-4e74-995e-53ce399f8b82","Type":"ContainerStarted","Data":"5777dda861dad04f87b52aec68064787a3e632a5bd399f797ee76b8510f778bd"} Dec 04 06:45:01 crc kubenswrapper[4685]: I1204 06:45:01.320113 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" event={"ID":"f322dc6e-91ac-4e74-995e-53ce399f8b82","Type":"ContainerStarted","Data":"faac2b560e6e3164af3ba5a35b3d3b80c3c47759eb055c3570653df809e520ee"} Dec 04 06:45:01 crc kubenswrapper[4685]: I1204 06:45:01.338672 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" podStartSLOduration=1.338655246 podStartE2EDuration="1.338655246s" podCreationTimestamp="2025-12-04 06:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 06:45:01.337617223 +0000 UTC m=+2038.525847998" watchObservedRunningTime="2025-12-04 06:45:01.338655246 +0000 UTC m=+2038.526886021" Dec 04 06:45:02 crc kubenswrapper[4685]: I1204 06:45:02.332003 4685 generic.go:334] "Generic (PLEG): container finished" podID="f322dc6e-91ac-4e74-995e-53ce399f8b82" containerID="5777dda861dad04f87b52aec68064787a3e632a5bd399f797ee76b8510f778bd" exitCode=0 Dec 04 06:45:02 crc kubenswrapper[4685]: I1204 06:45:02.332149 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" event={"ID":"f322dc6e-91ac-4e74-995e-53ce399f8b82","Type":"ContainerDied","Data":"5777dda861dad04f87b52aec68064787a3e632a5bd399f797ee76b8510f778bd"} Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.680014 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.739021 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9226\" (UniqueName: \"kubernetes.io/projected/f322dc6e-91ac-4e74-995e-53ce399f8b82-kube-api-access-m9226\") pod \"f322dc6e-91ac-4e74-995e-53ce399f8b82\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.739074 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f322dc6e-91ac-4e74-995e-53ce399f8b82-config-volume\") pod \"f322dc6e-91ac-4e74-995e-53ce399f8b82\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.739175 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f322dc6e-91ac-4e74-995e-53ce399f8b82-secret-volume\") pod \"f322dc6e-91ac-4e74-995e-53ce399f8b82\" (UID: \"f322dc6e-91ac-4e74-995e-53ce399f8b82\") " Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.739933 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f322dc6e-91ac-4e74-995e-53ce399f8b82-config-volume" (OuterVolumeSpecName: "config-volume") pod "f322dc6e-91ac-4e74-995e-53ce399f8b82" (UID: "f322dc6e-91ac-4e74-995e-53ce399f8b82"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.744849 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f322dc6e-91ac-4e74-995e-53ce399f8b82-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f322dc6e-91ac-4e74-995e-53ce399f8b82" (UID: "f322dc6e-91ac-4e74-995e-53ce399f8b82"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.744978 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f322dc6e-91ac-4e74-995e-53ce399f8b82-kube-api-access-m9226" (OuterVolumeSpecName: "kube-api-access-m9226") pod "f322dc6e-91ac-4e74-995e-53ce399f8b82" (UID: "f322dc6e-91ac-4e74-995e-53ce399f8b82"). InnerVolumeSpecName "kube-api-access-m9226". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.841975 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9226\" (UniqueName: \"kubernetes.io/projected/f322dc6e-91ac-4e74-995e-53ce399f8b82-kube-api-access-m9226\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.842034 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f322dc6e-91ac-4e74-995e-53ce399f8b82-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:03 crc kubenswrapper[4685]: I1204 06:45:03.842059 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f322dc6e-91ac-4e74-995e-53ce399f8b82-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:04 crc kubenswrapper[4685]: I1204 06:45:04.354763 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" event={"ID":"f322dc6e-91ac-4e74-995e-53ce399f8b82","Type":"ContainerDied","Data":"faac2b560e6e3164af3ba5a35b3d3b80c3c47759eb055c3570653df809e520ee"} Dec 04 06:45:04 crc kubenswrapper[4685]: I1204 06:45:04.355147 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faac2b560e6e3164af3ba5a35b3d3b80c3c47759eb055c3570653df809e520ee" Dec 04 06:45:04 crc kubenswrapper[4685]: I1204 06:45:04.354908 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz" Dec 04 06:45:04 crc kubenswrapper[4685]: I1204 06:45:04.452107 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm"] Dec 04 06:45:04 crc kubenswrapper[4685]: I1204 06:45:04.460479 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413800-slrlm"] Dec 04 06:45:05 crc kubenswrapper[4685]: I1204 06:45:05.137390 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc81c705-faab-44b6-8da9-96bee4bd043a" path="/var/lib/kubelet/pods/bc81c705-faab-44b6-8da9-96bee4bd043a/volumes" Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.121476 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.121871 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.121933 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.123491 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06d5f8fe304d520b079828f91b96cf934d5f4cad4ffa07e56214997b67c1a8a2"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.123621 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://06d5f8fe304d520b079828f91b96cf934d5f4cad4ffa07e56214997b67c1a8a2" gracePeriod=600 Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.416504 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="06d5f8fe304d520b079828f91b96cf934d5f4cad4ffa07e56214997b67c1a8a2" exitCode=0 Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.416565 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"06d5f8fe304d520b079828f91b96cf934d5f4cad4ffa07e56214997b67c1a8a2"} Dec 04 06:45:06 crc kubenswrapper[4685]: I1204 06:45:06.416683 4685 scope.go:117] "RemoveContainer" containerID="cf694e6700eebd3bc7b6566e2af48ffac3b3b88f976c0e9c692691dc9a2f3d47" Dec 04 06:45:07 crc kubenswrapper[4685]: I1204 06:45:07.429865 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8"} Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.086285 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-574x8"] Dec 04 06:45:16 crc kubenswrapper[4685]: E1204 06:45:16.087492 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f322dc6e-91ac-4e74-995e-53ce399f8b82" containerName="collect-profiles" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.087512 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f322dc6e-91ac-4e74-995e-53ce399f8b82" containerName="collect-profiles" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.087808 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f322dc6e-91ac-4e74-995e-53ce399f8b82" containerName="collect-profiles" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.089980 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.107386 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-574x8"] Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.216364 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns8tj\" (UniqueName: \"kubernetes.io/projected/016835d2-9e19-437e-8777-1e282451a48f-kube-api-access-ns8tj\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.216467 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-catalog-content\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.216539 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-utilities\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.319111 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns8tj\" (UniqueName: \"kubernetes.io/projected/016835d2-9e19-437e-8777-1e282451a48f-kube-api-access-ns8tj\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.319165 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-catalog-content\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.319191 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-utilities\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.319785 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-catalog-content\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.319964 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-utilities\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.353372 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns8tj\" (UniqueName: \"kubernetes.io/projected/016835d2-9e19-437e-8777-1e282451a48f-kube-api-access-ns8tj\") pod \"redhat-marketplace-574x8\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.427292 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:16 crc kubenswrapper[4685]: I1204 06:45:16.914617 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-574x8"] Dec 04 06:45:16 crc kubenswrapper[4685]: W1204 06:45:16.918514 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod016835d2_9e19_437e_8777_1e282451a48f.slice/crio-44b90f7c5968dba5d86313770a2899e7d48a0f98b81a66eb750177cf1f190d81 WatchSource:0}: Error finding container 44b90f7c5968dba5d86313770a2899e7d48a0f98b81a66eb750177cf1f190d81: Status 404 returned error can't find the container with id 44b90f7c5968dba5d86313770a2899e7d48a0f98b81a66eb750177cf1f190d81 Dec 04 06:45:17 crc kubenswrapper[4685]: I1204 06:45:17.532926 4685 generic.go:334] "Generic (PLEG): container finished" podID="016835d2-9e19-437e-8777-1e282451a48f" containerID="dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7" exitCode=0 Dec 04 06:45:17 crc kubenswrapper[4685]: I1204 06:45:17.532989 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerDied","Data":"dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7"} Dec 04 06:45:17 crc kubenswrapper[4685]: I1204 06:45:17.533205 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerStarted","Data":"44b90f7c5968dba5d86313770a2899e7d48a0f98b81a66eb750177cf1f190d81"} Dec 04 06:45:18 crc kubenswrapper[4685]: I1204 06:45:18.546454 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerStarted","Data":"bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f"} Dec 04 06:45:18 crc kubenswrapper[4685]: I1204 06:45:18.751971 4685 scope.go:117] "RemoveContainer" containerID="774b6a73a76a8a570017dfc9a3a611265d574edf851a53886a8fa38106e5e21a" Dec 04 06:45:19 crc kubenswrapper[4685]: I1204 06:45:19.562505 4685 generic.go:334] "Generic (PLEG): container finished" podID="016835d2-9e19-437e-8777-1e282451a48f" containerID="bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f" exitCode=0 Dec 04 06:45:19 crc kubenswrapper[4685]: I1204 06:45:19.562554 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerDied","Data":"bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f"} Dec 04 06:45:20 crc kubenswrapper[4685]: I1204 06:45:20.574734 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerStarted","Data":"df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782"} Dec 04 06:45:20 crc kubenswrapper[4685]: I1204 06:45:20.600350 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-574x8" podStartSLOduration=2.165022558 podStartE2EDuration="4.600327879s" podCreationTimestamp="2025-12-04 06:45:16 +0000 UTC" firstStartedPulling="2025-12-04 06:45:17.534639449 +0000 UTC m=+2054.722870244" lastFinishedPulling="2025-12-04 06:45:19.96994475 +0000 UTC m=+2057.158175565" observedRunningTime="2025-12-04 06:45:20.597932344 +0000 UTC m=+2057.786163149" watchObservedRunningTime="2025-12-04 06:45:20.600327879 +0000 UTC m=+2057.788558684" Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.427401 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.428222 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.478395 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.632134 4685 generic.go:334] "Generic (PLEG): container finished" podID="e75fdf41-c07d-4fdd-9abc-334e2164b96e" containerID="05fd69458eb3a713ec3aa4f8113fba05b5d9657289bec1bd9925b50803b4348b" exitCode=0 Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.632230 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" event={"ID":"e75fdf41-c07d-4fdd-9abc-334e2164b96e","Type":"ContainerDied","Data":"05fd69458eb3a713ec3aa4f8113fba05b5d9657289bec1bd9925b50803b4348b"} Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.687299 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:26 crc kubenswrapper[4685]: I1204 06:45:26.742993 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-574x8"] Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.158201 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.281903 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ovn-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.281969 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-nova-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282010 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-repo-setup-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282123 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282162 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ssh-key\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282200 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282309 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-neutron-metadata-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282443 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282522 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-inventory\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282630 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bscxq\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-kube-api-access-bscxq\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282693 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-telemetry-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282738 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282919 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-libvirt-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.282968 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-bootstrap-combined-ca-bundle\") pod \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\" (UID: \"e75fdf41-c07d-4fdd-9abc-334e2164b96e\") " Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.289575 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.290289 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.290706 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.291435 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.291496 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-kube-api-access-bscxq" (OuterVolumeSpecName: "kube-api-access-bscxq") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "kube-api-access-bscxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.291507 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.291868 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.293718 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.294870 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.295704 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.296129 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.296650 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.329047 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-inventory" (OuterVolumeSpecName: "inventory") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.330791 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e75fdf41-c07d-4fdd-9abc-334e2164b96e" (UID: "e75fdf41-c07d-4fdd-9abc-334e2164b96e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385211 4685 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385248 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385258 4685 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385267 4685 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385278 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385291 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385303 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385317 4685 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385330 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385340 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385436 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bscxq\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-kube-api-access-bscxq\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385475 4685 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385489 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e75fdf41-c07d-4fdd-9abc-334e2164b96e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.385500 4685 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fdf41-c07d-4fdd-9abc-334e2164b96e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.656522 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" event={"ID":"e75fdf41-c07d-4fdd-9abc-334e2164b96e","Type":"ContainerDied","Data":"7199c3c962ea0906458b439c52b5e44e5c806e425eb1f5a657913db7504bf9f8"} Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.657599 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7199c3c962ea0906458b439c52b5e44e5c806e425eb1f5a657913db7504bf9f8" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.656658 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-574x8" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="registry-server" containerID="cri-o://df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782" gracePeriod=2 Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.656557 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.820813 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f"] Dec 04 06:45:28 crc kubenswrapper[4685]: E1204 06:45:28.821498 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75fdf41-c07d-4fdd-9abc-334e2164b96e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.821614 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75fdf41-c07d-4fdd-9abc-334e2164b96e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.821962 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e75fdf41-c07d-4fdd-9abc-334e2164b96e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.822730 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.826908 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.827280 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.827540 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.827733 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.827959 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.828187 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f"] Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.897451 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.897603 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/252256a3-494e-45de-ae52-6f77715b468e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.897669 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.897720 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lvdg\" (UniqueName: \"kubernetes.io/projected/252256a3-494e-45de-ae52-6f77715b468e-kube-api-access-2lvdg\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.897864 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.999602 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.999645 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.999713 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/252256a3-494e-45de-ae52-6f77715b468e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.999745 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:28 crc kubenswrapper[4685]: I1204 06:45:28.999779 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lvdg\" (UniqueName: \"kubernetes.io/projected/252256a3-494e-45de-ae52-6f77715b468e-kube-api-access-2lvdg\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.001232 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/252256a3-494e-45de-ae52-6f77715b468e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.006767 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.008162 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.009904 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.021255 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lvdg\" (UniqueName: \"kubernetes.io/projected/252256a3-494e-45de-ae52-6f77715b468e-kube-api-access-2lvdg\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7nn2f\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.156841 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.160847 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.307161 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-utilities\") pod \"016835d2-9e19-437e-8777-1e282451a48f\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.307488 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-catalog-content\") pod \"016835d2-9e19-437e-8777-1e282451a48f\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.307681 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns8tj\" (UniqueName: \"kubernetes.io/projected/016835d2-9e19-437e-8777-1e282451a48f-kube-api-access-ns8tj\") pod \"016835d2-9e19-437e-8777-1e282451a48f\" (UID: \"016835d2-9e19-437e-8777-1e282451a48f\") " Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.308370 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-utilities" (OuterVolumeSpecName: "utilities") pod "016835d2-9e19-437e-8777-1e282451a48f" (UID: "016835d2-9e19-437e-8777-1e282451a48f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.320500 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/016835d2-9e19-437e-8777-1e282451a48f-kube-api-access-ns8tj" (OuterVolumeSpecName: "kube-api-access-ns8tj") pod "016835d2-9e19-437e-8777-1e282451a48f" (UID: "016835d2-9e19-437e-8777-1e282451a48f"). InnerVolumeSpecName "kube-api-access-ns8tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.330618 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "016835d2-9e19-437e-8777-1e282451a48f" (UID: "016835d2-9e19-437e-8777-1e282451a48f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.410161 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.410198 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016835d2-9e19-437e-8777-1e282451a48f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.410211 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns8tj\" (UniqueName: \"kubernetes.io/projected/016835d2-9e19-437e-8777-1e282451a48f-kube-api-access-ns8tj\") on node \"crc\" DevicePath \"\"" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.672461 4685 generic.go:334] "Generic (PLEG): container finished" podID="016835d2-9e19-437e-8777-1e282451a48f" containerID="df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782" exitCode=0 Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.672522 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-574x8" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.672532 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerDied","Data":"df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782"} Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.672691 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-574x8" event={"ID":"016835d2-9e19-437e-8777-1e282451a48f","Type":"ContainerDied","Data":"44b90f7c5968dba5d86313770a2899e7d48a0f98b81a66eb750177cf1f190d81"} Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.672714 4685 scope.go:117] "RemoveContainer" containerID="df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.698592 4685 scope.go:117] "RemoveContainer" containerID="bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.750150 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f"] Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.759842 4685 scope.go:117] "RemoveContainer" containerID="dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.766361 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-574x8"] Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.779106 4685 scope.go:117] "RemoveContainer" containerID="df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782" Dec 04 06:45:29 crc kubenswrapper[4685]: E1204 06:45:29.779551 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782\": container with ID starting with df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782 not found: ID does not exist" containerID="df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.779589 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782"} err="failed to get container status \"df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782\": rpc error: code = NotFound desc = could not find container \"df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782\": container with ID starting with df513509649e67c215961c980a266b002752625e86669fa7470deaa586fa3782 not found: ID does not exist" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.779616 4685 scope.go:117] "RemoveContainer" containerID="bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f" Dec 04 06:45:29 crc kubenswrapper[4685]: E1204 06:45:29.779918 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f\": container with ID starting with bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f not found: ID does not exist" containerID="bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.779944 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f"} err="failed to get container status \"bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f\": rpc error: code = NotFound desc = could not find container \"bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f\": container with ID starting with bf4518c363a6f8f9eedc55e908b541aad1b230edd825d800df688c120ac10b1f not found: ID does not exist" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.779960 4685 scope.go:117] "RemoveContainer" containerID="dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.780047 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-574x8"] Dec 04 06:45:29 crc kubenswrapper[4685]: E1204 06:45:29.780185 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7\": container with ID starting with dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7 not found: ID does not exist" containerID="dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7" Dec 04 06:45:29 crc kubenswrapper[4685]: I1204 06:45:29.780213 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7"} err="failed to get container status \"dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7\": rpc error: code = NotFound desc = could not find container \"dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7\": container with ID starting with dba63afd0d0a3985804b6deeffd8a6e4b2de8552542bdd35819ca1d8023b12c7 not found: ID does not exist" Dec 04 06:45:30 crc kubenswrapper[4685]: I1204 06:45:30.704231 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" event={"ID":"252256a3-494e-45de-ae52-6f77715b468e","Type":"ContainerStarted","Data":"2e73fc93002bca2ee2a62dffb98cd11f761891b53a8ecc5d4eb57fad1d2e3acf"} Dec 04 06:45:30 crc kubenswrapper[4685]: I1204 06:45:30.704601 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" event={"ID":"252256a3-494e-45de-ae52-6f77715b468e","Type":"ContainerStarted","Data":"72b6224b430256d2a938c25a4a84fbc410cc356a5367654da794d22a866b4404"} Dec 04 06:45:31 crc kubenswrapper[4685]: I1204 06:45:31.146002 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="016835d2-9e19-437e-8777-1e282451a48f" path="/var/lib/kubelet/pods/016835d2-9e19-437e-8777-1e282451a48f/volumes" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.147329 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" podStartSLOduration=34.936443872 podStartE2EDuration="35.147307425s" podCreationTimestamp="2025-12-04 06:45:28 +0000 UTC" firstStartedPulling="2025-12-04 06:45:29.736604301 +0000 UTC m=+2066.924835076" lastFinishedPulling="2025-12-04 06:45:29.947467844 +0000 UTC m=+2067.135698629" observedRunningTime="2025-12-04 06:45:30.724872952 +0000 UTC m=+2067.913103797" watchObservedRunningTime="2025-12-04 06:46:03.147307425 +0000 UTC m=+2100.335538200" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.163052 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-stpxp"] Dec 04 06:46:03 crc kubenswrapper[4685]: E1204 06:46:03.165035 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="extract-content" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.165064 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="extract-content" Dec 04 06:46:03 crc kubenswrapper[4685]: E1204 06:46:03.165094 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="extract-utilities" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.165103 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="extract-utilities" Dec 04 06:46:03 crc kubenswrapper[4685]: E1204 06:46:03.165125 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="registry-server" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.165133 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="registry-server" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.165441 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="016835d2-9e19-437e-8777-1e282451a48f" containerName="registry-server" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.167942 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.171368 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stpxp"] Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.209221 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-utilities\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.209300 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-catalog-content\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.209373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnrmz\" (UniqueName: \"kubernetes.io/projected/5890be7a-6c6b-4e03-9398-422a559b7452-kube-api-access-nnrmz\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.310915 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnrmz\" (UniqueName: \"kubernetes.io/projected/5890be7a-6c6b-4e03-9398-422a559b7452-kube-api-access-nnrmz\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.311350 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-utilities\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.311424 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-catalog-content\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.311948 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-utilities\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.311970 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-catalog-content\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.332544 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnrmz\" (UniqueName: \"kubernetes.io/projected/5890be7a-6c6b-4e03-9398-422a559b7452-kube-api-access-nnrmz\") pod \"community-operators-stpxp\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:03 crc kubenswrapper[4685]: I1204 06:46:03.510148 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:04 crc kubenswrapper[4685]: I1204 06:46:04.099886 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stpxp"] Dec 04 06:46:05 crc kubenswrapper[4685]: I1204 06:46:05.048509 4685 generic.go:334] "Generic (PLEG): container finished" podID="5890be7a-6c6b-4e03-9398-422a559b7452" containerID="f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45" exitCode=0 Dec 04 06:46:05 crc kubenswrapper[4685]: I1204 06:46:05.048596 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stpxp" event={"ID":"5890be7a-6c6b-4e03-9398-422a559b7452","Type":"ContainerDied","Data":"f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45"} Dec 04 06:46:05 crc kubenswrapper[4685]: I1204 06:46:05.051096 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stpxp" event={"ID":"5890be7a-6c6b-4e03-9398-422a559b7452","Type":"ContainerStarted","Data":"3b185661a7a01c4da3df9c7eea5ef3a1ab7a673355f6ccb0fe1e9d17d5d987ac"} Dec 04 06:46:05 crc kubenswrapper[4685]: I1204 06:46:05.051028 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:46:06 crc kubenswrapper[4685]: E1204 06:46:06.655792 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5890be7a_6c6b_4e03_9398_422a559b7452.slice/crio-986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778.scope\": RecentStats: unable to find data in memory cache]" Dec 04 06:46:07 crc kubenswrapper[4685]: I1204 06:46:07.079641 4685 generic.go:334] "Generic (PLEG): container finished" podID="5890be7a-6c6b-4e03-9398-422a559b7452" containerID="986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778" exitCode=0 Dec 04 06:46:07 crc kubenswrapper[4685]: I1204 06:46:07.079713 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stpxp" event={"ID":"5890be7a-6c6b-4e03-9398-422a559b7452","Type":"ContainerDied","Data":"986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778"} Dec 04 06:46:08 crc kubenswrapper[4685]: I1204 06:46:08.091385 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stpxp" event={"ID":"5890be7a-6c6b-4e03-9398-422a559b7452","Type":"ContainerStarted","Data":"d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a"} Dec 04 06:46:08 crc kubenswrapper[4685]: I1204 06:46:08.117745 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-stpxp" podStartSLOduration=2.596859003 podStartE2EDuration="5.117721117s" podCreationTimestamp="2025-12-04 06:46:03 +0000 UTC" firstStartedPulling="2025-12-04 06:46:05.050657114 +0000 UTC m=+2102.238887909" lastFinishedPulling="2025-12-04 06:46:07.571519238 +0000 UTC m=+2104.759750023" observedRunningTime="2025-12-04 06:46:08.113719521 +0000 UTC m=+2105.301950316" watchObservedRunningTime="2025-12-04 06:46:08.117721117 +0000 UTC m=+2105.305951932" Dec 04 06:46:13 crc kubenswrapper[4685]: I1204 06:46:13.510949 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:13 crc kubenswrapper[4685]: I1204 06:46:13.511495 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:13 crc kubenswrapper[4685]: I1204 06:46:13.594470 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:14 crc kubenswrapper[4685]: I1204 06:46:14.210662 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:14 crc kubenswrapper[4685]: I1204 06:46:14.271511 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-stpxp"] Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.176834 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-stpxp" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="registry-server" containerID="cri-o://d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a" gracePeriod=2 Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.651108 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.714237 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-utilities\") pod \"5890be7a-6c6b-4e03-9398-422a559b7452\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.714296 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnrmz\" (UniqueName: \"kubernetes.io/projected/5890be7a-6c6b-4e03-9398-422a559b7452-kube-api-access-nnrmz\") pod \"5890be7a-6c6b-4e03-9398-422a559b7452\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.714539 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-catalog-content\") pod \"5890be7a-6c6b-4e03-9398-422a559b7452\" (UID: \"5890be7a-6c6b-4e03-9398-422a559b7452\") " Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.715350 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-utilities" (OuterVolumeSpecName: "utilities") pod "5890be7a-6c6b-4e03-9398-422a559b7452" (UID: "5890be7a-6c6b-4e03-9398-422a559b7452"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.720751 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5890be7a-6c6b-4e03-9398-422a559b7452-kube-api-access-nnrmz" (OuterVolumeSpecName: "kube-api-access-nnrmz") pod "5890be7a-6c6b-4e03-9398-422a559b7452" (UID: "5890be7a-6c6b-4e03-9398-422a559b7452"). InnerVolumeSpecName "kube-api-access-nnrmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.774760 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5890be7a-6c6b-4e03-9398-422a559b7452" (UID: "5890be7a-6c6b-4e03-9398-422a559b7452"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.816988 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.817048 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnrmz\" (UniqueName: \"kubernetes.io/projected/5890be7a-6c6b-4e03-9398-422a559b7452-kube-api-access-nnrmz\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:16 crc kubenswrapper[4685]: I1204 06:46:16.817071 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5890be7a-6c6b-4e03-9398-422a559b7452-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.187745 4685 generic.go:334] "Generic (PLEG): container finished" podID="5890be7a-6c6b-4e03-9398-422a559b7452" containerID="d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a" exitCode=0 Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.187848 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stpxp" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.187853 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stpxp" event={"ID":"5890be7a-6c6b-4e03-9398-422a559b7452","Type":"ContainerDied","Data":"d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a"} Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.188206 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stpxp" event={"ID":"5890be7a-6c6b-4e03-9398-422a559b7452","Type":"ContainerDied","Data":"3b185661a7a01c4da3df9c7eea5ef3a1ab7a673355f6ccb0fe1e9d17d5d987ac"} Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.188227 4685 scope.go:117] "RemoveContainer" containerID="d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.235679 4685 scope.go:117] "RemoveContainer" containerID="986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.256488 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-stpxp"] Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.261901 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-stpxp"] Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.285825 4685 scope.go:117] "RemoveContainer" containerID="f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.347535 4685 scope.go:117] "RemoveContainer" containerID="d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a" Dec 04 06:46:17 crc kubenswrapper[4685]: E1204 06:46:17.351516 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a\": container with ID starting with d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a not found: ID does not exist" containerID="d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.351557 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a"} err="failed to get container status \"d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a\": rpc error: code = NotFound desc = could not find container \"d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a\": container with ID starting with d55842921170acb342c615df81963d81983943fef366031706662e3883aa387a not found: ID does not exist" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.351584 4685 scope.go:117] "RemoveContainer" containerID="986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778" Dec 04 06:46:17 crc kubenswrapper[4685]: E1204 06:46:17.355512 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778\": container with ID starting with 986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778 not found: ID does not exist" containerID="986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.355551 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778"} err="failed to get container status \"986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778\": rpc error: code = NotFound desc = could not find container \"986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778\": container with ID starting with 986a0c531a07513f051609127ae1b3803f9d6e895288e69fb4e063731bf5e778 not found: ID does not exist" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.355575 4685 scope.go:117] "RemoveContainer" containerID="f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45" Dec 04 06:46:17 crc kubenswrapper[4685]: E1204 06:46:17.360515 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45\": container with ID starting with f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45 not found: ID does not exist" containerID="f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45" Dec 04 06:46:17 crc kubenswrapper[4685]: I1204 06:46:17.360552 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45"} err="failed to get container status \"f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45\": rpc error: code = NotFound desc = could not find container \"f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45\": container with ID starting with f6ca09b6e3634202dae02ffc52c813448401e82a1dd155ffa3962649ed0cda45 not found: ID does not exist" Dec 04 06:46:19 crc kubenswrapper[4685]: I1204 06:46:19.139782 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" path="/var/lib/kubelet/pods/5890be7a-6c6b-4e03-9398-422a559b7452/volumes" Dec 04 06:46:44 crc kubenswrapper[4685]: I1204 06:46:44.467737 4685 generic.go:334] "Generic (PLEG): container finished" podID="252256a3-494e-45de-ae52-6f77715b468e" containerID="2e73fc93002bca2ee2a62dffb98cd11f761891b53a8ecc5d4eb57fad1d2e3acf" exitCode=0 Dec 04 06:46:44 crc kubenswrapper[4685]: I1204 06:46:44.467867 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" event={"ID":"252256a3-494e-45de-ae52-6f77715b468e","Type":"ContainerDied","Data":"2e73fc93002bca2ee2a62dffb98cd11f761891b53a8ecc5d4eb57fad1d2e3acf"} Dec 04 06:46:45 crc kubenswrapper[4685]: I1204 06:46:45.919868 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.027699 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lvdg\" (UniqueName: \"kubernetes.io/projected/252256a3-494e-45de-ae52-6f77715b468e-kube-api-access-2lvdg\") pod \"252256a3-494e-45de-ae52-6f77715b468e\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.027838 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/252256a3-494e-45de-ae52-6f77715b468e-ovncontroller-config-0\") pod \"252256a3-494e-45de-ae52-6f77715b468e\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.027903 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ovn-combined-ca-bundle\") pod \"252256a3-494e-45de-ae52-6f77715b468e\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.028109 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-inventory\") pod \"252256a3-494e-45de-ae52-6f77715b468e\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.028184 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ssh-key\") pod \"252256a3-494e-45de-ae52-6f77715b468e\" (UID: \"252256a3-494e-45de-ae52-6f77715b468e\") " Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.035768 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252256a3-494e-45de-ae52-6f77715b468e-kube-api-access-2lvdg" (OuterVolumeSpecName: "kube-api-access-2lvdg") pod "252256a3-494e-45de-ae52-6f77715b468e" (UID: "252256a3-494e-45de-ae52-6f77715b468e"). InnerVolumeSpecName "kube-api-access-2lvdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.036321 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "252256a3-494e-45de-ae52-6f77715b468e" (UID: "252256a3-494e-45de-ae52-6f77715b468e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.070552 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-inventory" (OuterVolumeSpecName: "inventory") pod "252256a3-494e-45de-ae52-6f77715b468e" (UID: "252256a3-494e-45de-ae52-6f77715b468e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.078589 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "252256a3-494e-45de-ae52-6f77715b468e" (UID: "252256a3-494e-45de-ae52-6f77715b468e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.087513 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252256a3-494e-45de-ae52-6f77715b468e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "252256a3-494e-45de-ae52-6f77715b468e" (UID: "252256a3-494e-45de-ae52-6f77715b468e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.132194 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.132229 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.132238 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lvdg\" (UniqueName: \"kubernetes.io/projected/252256a3-494e-45de-ae52-6f77715b468e-kube-api-access-2lvdg\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.132248 4685 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/252256a3-494e-45de-ae52-6f77715b468e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.132258 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/252256a3-494e-45de-ae52-6f77715b468e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.508130 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" event={"ID":"252256a3-494e-45de-ae52-6f77715b468e","Type":"ContainerDied","Data":"72b6224b430256d2a938c25a4a84fbc410cc356a5367654da794d22a866b4404"} Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.508603 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72b6224b430256d2a938c25a4a84fbc410cc356a5367654da794d22a866b4404" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.508696 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7nn2f" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.635357 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq"] Dec 04 06:46:46 crc kubenswrapper[4685]: E1204 06:46:46.635804 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="extract-utilities" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.635826 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="extract-utilities" Dec 04 06:46:46 crc kubenswrapper[4685]: E1204 06:46:46.635846 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="registry-server" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.635854 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="registry-server" Dec 04 06:46:46 crc kubenswrapper[4685]: E1204 06:46:46.635873 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="extract-content" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.635882 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="extract-content" Dec 04 06:46:46 crc kubenswrapper[4685]: E1204 06:46:46.635899 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252256a3-494e-45de-ae52-6f77715b468e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.635909 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="252256a3-494e-45de-ae52-6f77715b468e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.636141 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5890be7a-6c6b-4e03-9398-422a559b7452" containerName="registry-server" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.636184 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="252256a3-494e-45de-ae52-6f77715b468e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.636913 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.640598 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.640807 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.641024 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.641255 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.641507 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.642399 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.643376 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq"] Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.745991 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.746090 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.746146 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.746166 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.746445 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkfxf\" (UniqueName: \"kubernetes.io/projected/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-kube-api-access-qkfxf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.746645 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.849017 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.849266 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.849369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.849573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.849718 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkfxf\" (UniqueName: \"kubernetes.io/projected/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-kube-api-access-qkfxf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.849891 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.853821 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.854171 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.856574 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.861507 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.869850 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:46 crc kubenswrapper[4685]: I1204 06:46:46.882280 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkfxf\" (UniqueName: \"kubernetes.io/projected/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-kube-api-access-qkfxf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:47 crc kubenswrapper[4685]: I1204 06:46:47.009524 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:46:47 crc kubenswrapper[4685]: I1204 06:46:47.570641 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq"] Dec 04 06:46:47 crc kubenswrapper[4685]: W1204 06:46:47.572548 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb314b13b_4d0c_47f2_8fe8_1ab32ad74f28.slice/crio-887832db3522ea3a497ed173bc815a64d9849ffe0664b01f6f58ba152a436a6c WatchSource:0}: Error finding container 887832db3522ea3a497ed173bc815a64d9849ffe0664b01f6f58ba152a436a6c: Status 404 returned error can't find the container with id 887832db3522ea3a497ed173bc815a64d9849ffe0664b01f6f58ba152a436a6c Dec 04 06:46:48 crc kubenswrapper[4685]: I1204 06:46:48.551975 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" event={"ID":"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28","Type":"ContainerStarted","Data":"830f422ba4853a5c58d3b0ee71ab591107f87771533a7fa75064b7d8b0dc98a3"} Dec 04 06:46:48 crc kubenswrapper[4685]: I1204 06:46:48.552514 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" event={"ID":"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28","Type":"ContainerStarted","Data":"887832db3522ea3a497ed173bc815a64d9849ffe0664b01f6f58ba152a436a6c"} Dec 04 06:46:48 crc kubenswrapper[4685]: I1204 06:46:48.586142 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" podStartSLOduration=2.4153771600000002 podStartE2EDuration="2.586109779s" podCreationTimestamp="2025-12-04 06:46:46 +0000 UTC" firstStartedPulling="2025-12-04 06:46:47.574945234 +0000 UTC m=+2144.763176009" lastFinishedPulling="2025-12-04 06:46:47.745677853 +0000 UTC m=+2144.933908628" observedRunningTime="2025-12-04 06:46:48.575215657 +0000 UTC m=+2145.763446472" watchObservedRunningTime="2025-12-04 06:46:48.586109779 +0000 UTC m=+2145.774340584" Dec 04 06:47:06 crc kubenswrapper[4685]: I1204 06:47:06.121891 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:47:06 crc kubenswrapper[4685]: I1204 06:47:06.122538 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:47:26 crc kubenswrapper[4685]: I1204 06:47:26.861392 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8chjm"] Dec 04 06:47:26 crc kubenswrapper[4685]: I1204 06:47:26.864842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:26 crc kubenswrapper[4685]: I1204 06:47:26.916356 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8chjm"] Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.057074 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-utilities\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.057219 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jvbq\" (UniqueName: \"kubernetes.io/projected/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-kube-api-access-5jvbq\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.057688 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-catalog-content\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.159343 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-utilities\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.159399 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jvbq\" (UniqueName: \"kubernetes.io/projected/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-kube-api-access-5jvbq\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.159554 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-catalog-content\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.159973 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-utilities\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.160092 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-catalog-content\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.181565 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jvbq\" (UniqueName: \"kubernetes.io/projected/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-kube-api-access-5jvbq\") pod \"redhat-operators-8chjm\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.211570 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:27 crc kubenswrapper[4685]: I1204 06:47:27.675290 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8chjm"] Dec 04 06:47:28 crc kubenswrapper[4685]: I1204 06:47:28.006045 4685 generic.go:334] "Generic (PLEG): container finished" podID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerID="27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446" exitCode=0 Dec 04 06:47:28 crc kubenswrapper[4685]: I1204 06:47:28.006133 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerDied","Data":"27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446"} Dec 04 06:47:28 crc kubenswrapper[4685]: I1204 06:47:28.006399 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerStarted","Data":"149dba0ec4860c9a6edbaaffce697c59141f409bb086c2c0998aa46a4a2cf989"} Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.016545 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerStarted","Data":"15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88"} Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.271992 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-slpqj"] Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.277242 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.285974 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slpqj"] Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.396103 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-utilities\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.396593 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7bg\" (UniqueName: \"kubernetes.io/projected/1decb07e-f243-4be0-8185-7ed73cd949cb-kube-api-access-wf7bg\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.396615 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-catalog-content\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.498589 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7bg\" (UniqueName: \"kubernetes.io/projected/1decb07e-f243-4be0-8185-7ed73cd949cb-kube-api-access-wf7bg\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.498658 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-catalog-content\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.498722 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-utilities\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.499361 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-utilities\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.499667 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-catalog-content\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.530486 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7bg\" (UniqueName: \"kubernetes.io/projected/1decb07e-f243-4be0-8185-7ed73cd949cb-kube-api-access-wf7bg\") pod \"certified-operators-slpqj\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:29 crc kubenswrapper[4685]: I1204 06:47:29.620564 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:30 crc kubenswrapper[4685]: I1204 06:47:30.031128 4685 generic.go:334] "Generic (PLEG): container finished" podID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerID="15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88" exitCode=0 Dec 04 06:47:30 crc kubenswrapper[4685]: I1204 06:47:30.031365 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerDied","Data":"15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88"} Dec 04 06:47:30 crc kubenswrapper[4685]: I1204 06:47:30.091019 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slpqj"] Dec 04 06:47:30 crc kubenswrapper[4685]: W1204 06:47:30.091885 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1decb07e_f243_4be0_8185_7ed73cd949cb.slice/crio-010b0b53bcb8233b207b97eab7336c37c75917a36c599cb740f1c0eca2c4b70c WatchSource:0}: Error finding container 010b0b53bcb8233b207b97eab7336c37c75917a36c599cb740f1c0eca2c4b70c: Status 404 returned error can't find the container with id 010b0b53bcb8233b207b97eab7336c37c75917a36c599cb740f1c0eca2c4b70c Dec 04 06:47:31 crc kubenswrapper[4685]: I1204 06:47:31.044731 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerStarted","Data":"ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4"} Dec 04 06:47:31 crc kubenswrapper[4685]: I1204 06:47:31.047607 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerStarted","Data":"c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1"} Dec 04 06:47:31 crc kubenswrapper[4685]: I1204 06:47:31.047650 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerStarted","Data":"010b0b53bcb8233b207b97eab7336c37c75917a36c599cb740f1c0eca2c4b70c"} Dec 04 06:47:31 crc kubenswrapper[4685]: I1204 06:47:31.068510 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8chjm" podStartSLOduration=2.634708041 podStartE2EDuration="5.06849048s" podCreationTimestamp="2025-12-04 06:47:26 +0000 UTC" firstStartedPulling="2025-12-04 06:47:28.008070392 +0000 UTC m=+2185.196301167" lastFinishedPulling="2025-12-04 06:47:30.441852831 +0000 UTC m=+2187.630083606" observedRunningTime="2025-12-04 06:47:31.060157639 +0000 UTC m=+2188.248388424" watchObservedRunningTime="2025-12-04 06:47:31.06849048 +0000 UTC m=+2188.256721265" Dec 04 06:47:32 crc kubenswrapper[4685]: I1204 06:47:32.074532 4685 generic.go:334] "Generic (PLEG): container finished" podID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerID="c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1" exitCode=0 Dec 04 06:47:32 crc kubenswrapper[4685]: I1204 06:47:32.074597 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerDied","Data":"c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1"} Dec 04 06:47:33 crc kubenswrapper[4685]: I1204 06:47:33.086048 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerStarted","Data":"ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed"} Dec 04 06:47:34 crc kubenswrapper[4685]: I1204 06:47:34.095286 4685 generic.go:334] "Generic (PLEG): container finished" podID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerID="ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed" exitCode=0 Dec 04 06:47:34 crc kubenswrapper[4685]: I1204 06:47:34.095327 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerDied","Data":"ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed"} Dec 04 06:47:35 crc kubenswrapper[4685]: I1204 06:47:35.106274 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerStarted","Data":"5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa"} Dec 04 06:47:35 crc kubenswrapper[4685]: I1204 06:47:35.133140 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-slpqj" podStartSLOduration=3.640217674 podStartE2EDuration="6.133125745s" podCreationTimestamp="2025-12-04 06:47:29 +0000 UTC" firstStartedPulling="2025-12-04 06:47:32.07791147 +0000 UTC m=+2189.266142265" lastFinishedPulling="2025-12-04 06:47:34.570819551 +0000 UTC m=+2191.759050336" observedRunningTime="2025-12-04 06:47:35.1291428 +0000 UTC m=+2192.317373585" watchObservedRunningTime="2025-12-04 06:47:35.133125745 +0000 UTC m=+2192.321356520" Dec 04 06:47:36 crc kubenswrapper[4685]: I1204 06:47:36.121267 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:47:36 crc kubenswrapper[4685]: I1204 06:47:36.121599 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:47:37 crc kubenswrapper[4685]: I1204 06:47:37.212577 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:37 crc kubenswrapper[4685]: I1204 06:47:37.213082 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:37 crc kubenswrapper[4685]: I1204 06:47:37.295176 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:38 crc kubenswrapper[4685]: I1204 06:47:38.209501 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:38 crc kubenswrapper[4685]: I1204 06:47:38.450256 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8chjm"] Dec 04 06:47:39 crc kubenswrapper[4685]: I1204 06:47:39.621517 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:39 crc kubenswrapper[4685]: I1204 06:47:39.622882 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:39 crc kubenswrapper[4685]: I1204 06:47:39.686742 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.186374 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8chjm" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="registry-server" containerID="cri-o://ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4" gracePeriod=2 Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.258711 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.852279 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-slpqj"] Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.876384 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.964992 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jvbq\" (UniqueName: \"kubernetes.io/projected/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-kube-api-access-5jvbq\") pod \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.965061 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-catalog-content\") pod \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.965131 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-utilities\") pod \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\" (UID: \"b76c9b3d-fb69-470a-bbc9-5b7a82d69108\") " Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.966573 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-utilities" (OuterVolumeSpecName: "utilities") pod "b76c9b3d-fb69-470a-bbc9-5b7a82d69108" (UID: "b76c9b3d-fb69-470a-bbc9-5b7a82d69108"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:47:40 crc kubenswrapper[4685]: I1204 06:47:40.977153 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-kube-api-access-5jvbq" (OuterVolumeSpecName: "kube-api-access-5jvbq") pod "b76c9b3d-fb69-470a-bbc9-5b7a82d69108" (UID: "b76c9b3d-fb69-470a-bbc9-5b7a82d69108"). InnerVolumeSpecName "kube-api-access-5jvbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.067218 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.067251 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jvbq\" (UniqueName: \"kubernetes.io/projected/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-kube-api-access-5jvbq\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.092319 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b76c9b3d-fb69-470a-bbc9-5b7a82d69108" (UID: "b76c9b3d-fb69-470a-bbc9-5b7a82d69108"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.170979 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76c9b3d-fb69-470a-bbc9-5b7a82d69108-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.196448 4685 generic.go:334] "Generic (PLEG): container finished" podID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerID="ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4" exitCode=0 Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.196506 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8chjm" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.196509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerDied","Data":"ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4"} Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.196583 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8chjm" event={"ID":"b76c9b3d-fb69-470a-bbc9-5b7a82d69108","Type":"ContainerDied","Data":"149dba0ec4860c9a6edbaaffce697c59141f409bb086c2c0998aa46a4a2cf989"} Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.196611 4685 scope.go:117] "RemoveContainer" containerID="ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.225533 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8chjm"] Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.227460 4685 scope.go:117] "RemoveContainer" containerID="15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.236807 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8chjm"] Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.254059 4685 scope.go:117] "RemoveContainer" containerID="27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.290837 4685 scope.go:117] "RemoveContainer" containerID="ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4" Dec 04 06:47:41 crc kubenswrapper[4685]: E1204 06:47:41.291618 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4\": container with ID starting with ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4 not found: ID does not exist" containerID="ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.291672 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4"} err="failed to get container status \"ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4\": rpc error: code = NotFound desc = could not find container \"ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4\": container with ID starting with ffd2884052daaef7a6cd814d68c55d35ab1293de122c90e7d6bf9442be6ce7b4 not found: ID does not exist" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.291697 4685 scope.go:117] "RemoveContainer" containerID="15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88" Dec 04 06:47:41 crc kubenswrapper[4685]: E1204 06:47:41.292068 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88\": container with ID starting with 15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88 not found: ID does not exist" containerID="15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.292118 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88"} err="failed to get container status \"15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88\": rpc error: code = NotFound desc = could not find container \"15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88\": container with ID starting with 15ffdfee1e7d5027a2b77de6ae92542d2f88f8f9cd83838eb3b462dadc844c88 not found: ID does not exist" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.292146 4685 scope.go:117] "RemoveContainer" containerID="27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446" Dec 04 06:47:41 crc kubenswrapper[4685]: E1204 06:47:41.292465 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446\": container with ID starting with 27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446 not found: ID does not exist" containerID="27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446" Dec 04 06:47:41 crc kubenswrapper[4685]: I1204 06:47:41.292505 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446"} err="failed to get container status \"27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446\": rpc error: code = NotFound desc = could not find container \"27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446\": container with ID starting with 27d0720ce1efb41e83cc21eefa0ee180ccd37f0cd6ce549f379c42e9708f9446 not found: ID does not exist" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.210371 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-slpqj" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="registry-server" containerID="cri-o://5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa" gracePeriod=2 Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.727054 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.806999 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-catalog-content\") pod \"1decb07e-f243-4be0-8185-7ed73cd949cb\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.807116 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-utilities\") pod \"1decb07e-f243-4be0-8185-7ed73cd949cb\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.807196 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf7bg\" (UniqueName: \"kubernetes.io/projected/1decb07e-f243-4be0-8185-7ed73cd949cb-kube-api-access-wf7bg\") pod \"1decb07e-f243-4be0-8185-7ed73cd949cb\" (UID: \"1decb07e-f243-4be0-8185-7ed73cd949cb\") " Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.809393 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-utilities" (OuterVolumeSpecName: "utilities") pod "1decb07e-f243-4be0-8185-7ed73cd949cb" (UID: "1decb07e-f243-4be0-8185-7ed73cd949cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.831591 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1decb07e-f243-4be0-8185-7ed73cd949cb-kube-api-access-wf7bg" (OuterVolumeSpecName: "kube-api-access-wf7bg") pod "1decb07e-f243-4be0-8185-7ed73cd949cb" (UID: "1decb07e-f243-4be0-8185-7ed73cd949cb"). InnerVolumeSpecName "kube-api-access-wf7bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.881584 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1decb07e-f243-4be0-8185-7ed73cd949cb" (UID: "1decb07e-f243-4be0-8185-7ed73cd949cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.909293 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.909323 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1decb07e-f243-4be0-8185-7ed73cd949cb-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:42 crc kubenswrapper[4685]: I1204 06:47:42.909334 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf7bg\" (UniqueName: \"kubernetes.io/projected/1decb07e-f243-4be0-8185-7ed73cd949cb-kube-api-access-wf7bg\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.144060 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" path="/var/lib/kubelet/pods/b76c9b3d-fb69-470a-bbc9-5b7a82d69108/volumes" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.229158 4685 generic.go:334] "Generic (PLEG): container finished" podID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerID="5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa" exitCode=0 Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.229213 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerDied","Data":"5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa"} Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.229230 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slpqj" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.229304 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slpqj" event={"ID":"1decb07e-f243-4be0-8185-7ed73cd949cb","Type":"ContainerDied","Data":"010b0b53bcb8233b207b97eab7336c37c75917a36c599cb740f1c0eca2c4b70c"} Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.229343 4685 scope.go:117] "RemoveContainer" containerID="5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.262940 4685 scope.go:117] "RemoveContainer" containerID="ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.269274 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-slpqj"] Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.296523 4685 scope.go:117] "RemoveContainer" containerID="c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.302350 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-slpqj"] Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.329617 4685 scope.go:117] "RemoveContainer" containerID="5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa" Dec 04 06:47:43 crc kubenswrapper[4685]: E1204 06:47:43.331401 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa\": container with ID starting with 5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa not found: ID does not exist" containerID="5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.331465 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa"} err="failed to get container status \"5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa\": rpc error: code = NotFound desc = could not find container \"5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa\": container with ID starting with 5fc803177096991109918c012678c2dcf88b638ae9d48ee5aa85529f3e79cffa not found: ID does not exist" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.331491 4685 scope.go:117] "RemoveContainer" containerID="ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed" Dec 04 06:47:43 crc kubenswrapper[4685]: E1204 06:47:43.331852 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed\": container with ID starting with ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed not found: ID does not exist" containerID="ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.331938 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed"} err="failed to get container status \"ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed\": rpc error: code = NotFound desc = could not find container \"ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed\": container with ID starting with ea1d57823923dea48beee1670669853bd2540416999038091994e05030d235ed not found: ID does not exist" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.332008 4685 scope.go:117] "RemoveContainer" containerID="c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1" Dec 04 06:47:43 crc kubenswrapper[4685]: E1204 06:47:43.332492 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1\": container with ID starting with c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1 not found: ID does not exist" containerID="c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1" Dec 04 06:47:43 crc kubenswrapper[4685]: I1204 06:47:43.332584 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1"} err="failed to get container status \"c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1\": rpc error: code = NotFound desc = could not find container \"c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1\": container with ID starting with c58002bc4bf0895f416c8478444e79da0e32a4e8e49d67031b26d46dc056ded1 not found: ID does not exist" Dec 04 06:47:45 crc kubenswrapper[4685]: I1204 06:47:45.139487 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" path="/var/lib/kubelet/pods/1decb07e-f243-4be0-8185-7ed73cd949cb/volumes" Dec 04 06:47:47 crc kubenswrapper[4685]: I1204 06:47:47.275221 4685 generic.go:334] "Generic (PLEG): container finished" podID="b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" containerID="830f422ba4853a5c58d3b0ee71ab591107f87771533a7fa75064b7d8b0dc98a3" exitCode=0 Dec 04 06:47:47 crc kubenswrapper[4685]: I1204 06:47:47.275322 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" event={"ID":"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28","Type":"ContainerDied","Data":"830f422ba4853a5c58d3b0ee71ab591107f87771533a7fa75064b7d8b0dc98a3"} Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.750911 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.924670 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-ovn-metadata-agent-neutron-config-0\") pod \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.924889 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkfxf\" (UniqueName: \"kubernetes.io/projected/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-kube-api-access-qkfxf\") pod \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.926123 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-metadata-combined-ca-bundle\") pod \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.926178 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-nova-metadata-neutron-config-0\") pod \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.926217 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-ssh-key\") pod \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.926257 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-inventory\") pod \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\" (UID: \"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28\") " Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.933194 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-kube-api-access-qkfxf" (OuterVolumeSpecName: "kube-api-access-qkfxf") pod "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" (UID: "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28"). InnerVolumeSpecName "kube-api-access-qkfxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.935912 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" (UID: "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.968744 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-inventory" (OuterVolumeSpecName: "inventory") pod "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" (UID: "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.978659 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" (UID: "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:47:48 crc kubenswrapper[4685]: I1204 06:47:48.983290 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" (UID: "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.002601 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" (UID: "b314b13b-4d0c-47f2-8fe8-1ab32ad74f28"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.029049 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkfxf\" (UniqueName: \"kubernetes.io/projected/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-kube-api-access-qkfxf\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.029079 4685 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.029089 4685 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.029098 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.029108 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.029116 4685 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b314b13b-4d0c-47f2-8fe8-1ab32ad74f28-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.296654 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" event={"ID":"b314b13b-4d0c-47f2-8fe8-1ab32ad74f28","Type":"ContainerDied","Data":"887832db3522ea3a497ed173bc815a64d9849ffe0664b01f6f58ba152a436a6c"} Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.296706 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="887832db3522ea3a497ed173bc815a64d9849ffe0664b01f6f58ba152a436a6c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.297531 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429184 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c"] Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429593 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="registry-server" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429608 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="registry-server" Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429628 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="extract-content" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429636 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="extract-content" Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429684 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429692 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429707 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="registry-server" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429712 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="registry-server" Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429723 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="extract-utilities" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429732 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="extract-utilities" Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429751 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="extract-utilities" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429758 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="extract-utilities" Dec 04 06:47:49 crc kubenswrapper[4685]: E1204 06:47:49.429770 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="extract-content" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429776 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="extract-content" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429952 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1decb07e-f243-4be0-8185-7ed73cd949cb" containerName="registry-server" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429963 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b76c9b3d-fb69-470a-bbc9-5b7a82d69108" containerName="registry-server" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.429976 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b314b13b-4d0c-47f2-8fe8-1ab32ad74f28" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.430555 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.433952 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.434245 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.434738 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.436911 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.437229 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.440138 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c"] Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.538324 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.538515 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.538650 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.538706 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cljlh\" (UniqueName: \"kubernetes.io/projected/5ae85205-468a-4d79-855f-0efa0c8f3534-kube-api-access-cljlh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.538727 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.640651 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.640832 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.640989 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cljlh\" (UniqueName: \"kubernetes.io/projected/5ae85205-468a-4d79-855f-0efa0c8f3534-kube-api-access-cljlh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.641045 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.641189 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.646132 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.646580 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.647164 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.651271 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.658840 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cljlh\" (UniqueName: \"kubernetes.io/projected/5ae85205-468a-4d79-855f-0efa0c8f3534-kube-api-access-cljlh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:49 crc kubenswrapper[4685]: I1204 06:47:49.761020 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:47:50 crc kubenswrapper[4685]: I1204 06:47:50.320769 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c"] Dec 04 06:47:51 crc kubenswrapper[4685]: I1204 06:47:51.319401 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" event={"ID":"5ae85205-468a-4d79-855f-0efa0c8f3534","Type":"ContainerStarted","Data":"c92481aa27710a4723999bfcfc2288cbf60690842ada3d1e1c2d1ddce415f29c"} Dec 04 06:47:51 crc kubenswrapper[4685]: I1204 06:47:51.319758 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" event={"ID":"5ae85205-468a-4d79-855f-0efa0c8f3534","Type":"ContainerStarted","Data":"c33bdc206b337af44b4a6b7d56d3a59ab99b2a891b6df74f75e972dd8d94cc57"} Dec 04 06:47:51 crc kubenswrapper[4685]: I1204 06:47:51.347503 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" podStartSLOduration=2.136055107 podStartE2EDuration="2.347480701s" podCreationTimestamp="2025-12-04 06:47:49 +0000 UTC" firstStartedPulling="2025-12-04 06:47:50.364009473 +0000 UTC m=+2207.552240248" lastFinishedPulling="2025-12-04 06:47:50.575435057 +0000 UTC m=+2207.763665842" observedRunningTime="2025-12-04 06:47:51.337319673 +0000 UTC m=+2208.525550448" watchObservedRunningTime="2025-12-04 06:47:51.347480701 +0000 UTC m=+2208.535711486" Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.121563 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.122292 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.122358 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.123382 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.123499 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" gracePeriod=600 Dec 04 06:48:06 crc kubenswrapper[4685]: E1204 06:48:06.249063 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.476757 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" exitCode=0 Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.476826 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8"} Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.476931 4685 scope.go:117] "RemoveContainer" containerID="06d5f8fe304d520b079828f91b96cf934d5f4cad4ffa07e56214997b67c1a8a2" Dec 04 06:48:06 crc kubenswrapper[4685]: I1204 06:48:06.477734 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:48:06 crc kubenswrapper[4685]: E1204 06:48:06.478166 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:48:18 crc kubenswrapper[4685]: I1204 06:48:18.125477 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:48:18 crc kubenswrapper[4685]: E1204 06:48:18.126356 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:48:29 crc kubenswrapper[4685]: I1204 06:48:29.126540 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:48:29 crc kubenswrapper[4685]: E1204 06:48:29.127304 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:48:40 crc kubenswrapper[4685]: I1204 06:48:40.125897 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:48:40 crc kubenswrapper[4685]: E1204 06:48:40.127187 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:48:55 crc kubenswrapper[4685]: I1204 06:48:55.126659 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:48:55 crc kubenswrapper[4685]: E1204 06:48:55.127890 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:49:06 crc kubenswrapper[4685]: I1204 06:49:06.125337 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:49:06 crc kubenswrapper[4685]: E1204 06:49:06.126008 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:49:21 crc kubenswrapper[4685]: I1204 06:49:21.126289 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:49:21 crc kubenswrapper[4685]: E1204 06:49:21.127551 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:49:36 crc kubenswrapper[4685]: I1204 06:49:36.125639 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:49:36 crc kubenswrapper[4685]: E1204 06:49:36.126696 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:49:48 crc kubenswrapper[4685]: I1204 06:49:48.126121 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:49:48 crc kubenswrapper[4685]: E1204 06:49:48.126913 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:50:02 crc kubenswrapper[4685]: I1204 06:50:02.125918 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:50:02 crc kubenswrapper[4685]: E1204 06:50:02.126647 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:50:14 crc kubenswrapper[4685]: I1204 06:50:14.125270 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:50:14 crc kubenswrapper[4685]: E1204 06:50:14.126292 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:50:26 crc kubenswrapper[4685]: I1204 06:50:26.125937 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:50:26 crc kubenswrapper[4685]: E1204 06:50:26.126798 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:50:40 crc kubenswrapper[4685]: I1204 06:50:40.125594 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:50:40 crc kubenswrapper[4685]: E1204 06:50:40.126499 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:50:53 crc kubenswrapper[4685]: I1204 06:50:53.132025 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:50:53 crc kubenswrapper[4685]: E1204 06:50:53.133104 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:51:06 crc kubenswrapper[4685]: I1204 06:51:06.125989 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:51:06 crc kubenswrapper[4685]: E1204 06:51:06.126859 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:51:18 crc kubenswrapper[4685]: I1204 06:51:18.126722 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:51:18 crc kubenswrapper[4685]: E1204 06:51:18.129269 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:51:31 crc kubenswrapper[4685]: I1204 06:51:31.126628 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:51:31 crc kubenswrapper[4685]: E1204 06:51:31.127644 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:51:44 crc kubenswrapper[4685]: I1204 06:51:44.126241 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:51:44 crc kubenswrapper[4685]: E1204 06:51:44.127530 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:51:58 crc kubenswrapper[4685]: I1204 06:51:58.126138 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:51:58 crc kubenswrapper[4685]: E1204 06:51:58.128436 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:52:13 crc kubenswrapper[4685]: I1204 06:52:13.131371 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:52:13 crc kubenswrapper[4685]: E1204 06:52:13.132232 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:52:25 crc kubenswrapper[4685]: I1204 06:52:25.125744 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:52:25 crc kubenswrapper[4685]: E1204 06:52:25.126903 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:52:30 crc kubenswrapper[4685]: I1204 06:52:30.234779 4685 generic.go:334] "Generic (PLEG): container finished" podID="5ae85205-468a-4d79-855f-0efa0c8f3534" containerID="c92481aa27710a4723999bfcfc2288cbf60690842ada3d1e1c2d1ddce415f29c" exitCode=0 Dec 04 06:52:30 crc kubenswrapper[4685]: I1204 06:52:30.234908 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" event={"ID":"5ae85205-468a-4d79-855f-0efa0c8f3534","Type":"ContainerDied","Data":"c92481aa27710a4723999bfcfc2288cbf60690842ada3d1e1c2d1ddce415f29c"} Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.785121 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.907370 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-inventory\") pod \"5ae85205-468a-4d79-855f-0efa0c8f3534\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.907493 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-secret-0\") pod \"5ae85205-468a-4d79-855f-0efa0c8f3534\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.907577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-combined-ca-bundle\") pod \"5ae85205-468a-4d79-855f-0efa0c8f3534\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.907668 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-ssh-key\") pod \"5ae85205-468a-4d79-855f-0efa0c8f3534\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.907707 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cljlh\" (UniqueName: \"kubernetes.io/projected/5ae85205-468a-4d79-855f-0efa0c8f3534-kube-api-access-cljlh\") pod \"5ae85205-468a-4d79-855f-0efa0c8f3534\" (UID: \"5ae85205-468a-4d79-855f-0efa0c8f3534\") " Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.914821 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae85205-468a-4d79-855f-0efa0c8f3534-kube-api-access-cljlh" (OuterVolumeSpecName: "kube-api-access-cljlh") pod "5ae85205-468a-4d79-855f-0efa0c8f3534" (UID: "5ae85205-468a-4d79-855f-0efa0c8f3534"). InnerVolumeSpecName "kube-api-access-cljlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.915855 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5ae85205-468a-4d79-855f-0efa0c8f3534" (UID: "5ae85205-468a-4d79-855f-0efa0c8f3534"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.934163 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-inventory" (OuterVolumeSpecName: "inventory") pod "5ae85205-468a-4d79-855f-0efa0c8f3534" (UID: "5ae85205-468a-4d79-855f-0efa0c8f3534"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.936893 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "5ae85205-468a-4d79-855f-0efa0c8f3534" (UID: "5ae85205-468a-4d79-855f-0efa0c8f3534"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:52:31 crc kubenswrapper[4685]: I1204 06:52:31.938989 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5ae85205-468a-4d79-855f-0efa0c8f3534" (UID: "5ae85205-468a-4d79-855f-0efa0c8f3534"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.009938 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.009983 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cljlh\" (UniqueName: \"kubernetes.io/projected/5ae85205-468a-4d79-855f-0efa0c8f3534-kube-api-access-cljlh\") on node \"crc\" DevicePath \"\"" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.009993 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.010002 4685 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.010014 4685 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae85205-468a-4d79-855f-0efa0c8f3534-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.254943 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" event={"ID":"5ae85205-468a-4d79-855f-0efa0c8f3534","Type":"ContainerDied","Data":"c33bdc206b337af44b4a6b7d56d3a59ab99b2a891b6df74f75e972dd8d94cc57"} Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.254989 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c33bdc206b337af44b4a6b7d56d3a59ab99b2a891b6df74f75e972dd8d94cc57" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.254994 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.346333 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc"] Dec 04 06:52:32 crc kubenswrapper[4685]: E1204 06:52:32.346953 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae85205-468a-4d79-855f-0efa0c8f3534" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.347043 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae85205-468a-4d79-855f-0efa0c8f3534" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.347368 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae85205-468a-4d79-855f-0efa0c8f3534" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.348076 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.350146 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.351639 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.352276 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.352441 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.352541 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.352640 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.365752 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.366386 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc"] Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.417001 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.417744 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.417881 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.418002 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.418142 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.418216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.418348 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh52k\" (UniqueName: \"kubernetes.io/projected/1d8d290a-eebc-4783-be38-9b592fa66f25-kube-api-access-lh52k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.418554 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.418710 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.520983 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521047 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521157 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521228 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521263 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521327 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh52k\" (UniqueName: \"kubernetes.io/projected/1d8d290a-eebc-4783-be38-9b592fa66f25-kube-api-access-lh52k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521462 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.521521 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.523143 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.527322 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.528122 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.528366 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.530829 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.532051 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.534206 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.536061 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.539582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh52k\" (UniqueName: \"kubernetes.io/projected/1d8d290a-eebc-4783-be38-9b592fa66f25-kube-api-access-lh52k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhdpc\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:32 crc kubenswrapper[4685]: I1204 06:52:32.684650 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:52:33 crc kubenswrapper[4685]: I1204 06:52:33.201111 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc"] Dec 04 06:52:33 crc kubenswrapper[4685]: I1204 06:52:33.207542 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:52:33 crc kubenswrapper[4685]: I1204 06:52:33.262890 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" event={"ID":"1d8d290a-eebc-4783-be38-9b592fa66f25","Type":"ContainerStarted","Data":"7118043765ec657f9c0357a7c3c56bfe0366bfffb25b89647758cd6c2725dd3c"} Dec 04 06:52:34 crc kubenswrapper[4685]: I1204 06:52:34.273065 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" event={"ID":"1d8d290a-eebc-4783-be38-9b592fa66f25","Type":"ContainerStarted","Data":"77aee54fffd3e79e6a858b451e75ce48fe22d2a47dc3ac278ed4882059ec2898"} Dec 04 06:52:34 crc kubenswrapper[4685]: I1204 06:52:34.295110 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" podStartSLOduration=2.123176004 podStartE2EDuration="2.295090641s" podCreationTimestamp="2025-12-04 06:52:32 +0000 UTC" firstStartedPulling="2025-12-04 06:52:33.207311687 +0000 UTC m=+2490.395542452" lastFinishedPulling="2025-12-04 06:52:33.379226314 +0000 UTC m=+2490.567457089" observedRunningTime="2025-12-04 06:52:34.289197966 +0000 UTC m=+2491.477428751" watchObservedRunningTime="2025-12-04 06:52:34.295090641 +0000 UTC m=+2491.483321416" Dec 04 06:52:40 crc kubenswrapper[4685]: I1204 06:52:40.126276 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:52:40 crc kubenswrapper[4685]: E1204 06:52:40.127087 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:52:55 crc kubenswrapper[4685]: I1204 06:52:55.126519 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:52:55 crc kubenswrapper[4685]: E1204 06:52:55.127298 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:53:09 crc kubenswrapper[4685]: I1204 06:53:09.125977 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:53:09 crc kubenswrapper[4685]: I1204 06:53:09.614010 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"494f3cc360254283c3eea2f0c461cdde3ec1104223dd2529efff313cf274bc96"} Dec 04 06:55:36 crc kubenswrapper[4685]: I1204 06:55:36.121459 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:55:36 crc kubenswrapper[4685]: I1204 06:55:36.122305 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:55:40 crc kubenswrapper[4685]: I1204 06:55:40.677891 4685 generic.go:334] "Generic (PLEG): container finished" podID="1d8d290a-eebc-4783-be38-9b592fa66f25" containerID="77aee54fffd3e79e6a858b451e75ce48fe22d2a47dc3ac278ed4882059ec2898" exitCode=0 Dec 04 06:55:40 crc kubenswrapper[4685]: I1204 06:55:40.678600 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" event={"ID":"1d8d290a-eebc-4783-be38-9b592fa66f25","Type":"ContainerDied","Data":"77aee54fffd3e79e6a858b451e75ce48fe22d2a47dc3ac278ed4882059ec2898"} Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.098054 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.267354 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-0\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.267760 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh52k\" (UniqueName: \"kubernetes.io/projected/1d8d290a-eebc-4783-be38-9b592fa66f25-kube-api-access-lh52k\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.267817 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-1\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.267859 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-inventory\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.267915 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-combined-ca-bundle\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.267966 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-ssh-key\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.268092 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-extra-config-0\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.268149 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-1\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.268193 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-0\") pod \"1d8d290a-eebc-4783-be38-9b592fa66f25\" (UID: \"1d8d290a-eebc-4783-be38-9b592fa66f25\") " Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.275927 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8d290a-eebc-4783-be38-9b592fa66f25-kube-api-access-lh52k" (OuterVolumeSpecName: "kube-api-access-lh52k") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "kube-api-access-lh52k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.276918 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.302619 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.304782 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.308855 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.309527 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.321210 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-inventory" (OuterVolumeSpecName: "inventory") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.321768 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.349604 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "1d8d290a-eebc-4783-be38-9b592fa66f25" (UID: "1d8d290a-eebc-4783-be38-9b592fa66f25"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.370959 4685 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.370999 4685 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371016 4685 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371028 4685 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371041 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh52k\" (UniqueName: \"kubernetes.io/projected/1d8d290a-eebc-4783-be38-9b592fa66f25-kube-api-access-lh52k\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371054 4685 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371066 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371077 4685 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.371100 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8d290a-eebc-4783-be38-9b592fa66f25-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.698242 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" event={"ID":"1d8d290a-eebc-4783-be38-9b592fa66f25","Type":"ContainerDied","Data":"7118043765ec657f9c0357a7c3c56bfe0366bfffb25b89647758cd6c2725dd3c"} Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.698581 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7118043765ec657f9c0357a7c3c56bfe0366bfffb25b89647758cd6c2725dd3c" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.698279 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhdpc" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.789182 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4"] Dec 04 06:55:42 crc kubenswrapper[4685]: E1204 06:55:42.790882 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8d290a-eebc-4783-be38-9b592fa66f25" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.790900 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8d290a-eebc-4783-be38-9b592fa66f25" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.791113 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8d290a-eebc-4783-be38-9b592fa66f25" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.791904 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.798744 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4"] Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.848927 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.849002 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.849237 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wcjht" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.849267 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.849954 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.984650 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.984746 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.984815 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/175a4186-effa-431c-9808-33f3fa73119b-kube-api-access-d58r4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.984944 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.985266 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.985402 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:42 crc kubenswrapper[4685]: I1204 06:55:42.985575 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087200 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087282 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/175a4186-effa-431c-9808-33f3fa73119b-kube-api-access-d58r4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087323 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087431 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087484 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087540 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.087571 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.092457 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.092660 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.093617 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.094829 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.095207 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.095655 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.106169 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/175a4186-effa-431c-9808-33f3fa73119b-kube-api-access-d58r4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.162725 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:55:43 crc kubenswrapper[4685]: I1204 06:55:43.742857 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4"] Dec 04 06:55:44 crc kubenswrapper[4685]: I1204 06:55:44.715506 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" event={"ID":"175a4186-effa-431c-9808-33f3fa73119b","Type":"ContainerStarted","Data":"8a8f73b62b67382a4801a0a47fbbbe4081155a86a43488323a9c651036f0cbad"} Dec 04 06:55:44 crc kubenswrapper[4685]: I1204 06:55:44.715899 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" event={"ID":"175a4186-effa-431c-9808-33f3fa73119b","Type":"ContainerStarted","Data":"baef2a430f8266483c16cace0782ad5784176214c49fae97f1f33f8cc539c356"} Dec 04 06:55:44 crc kubenswrapper[4685]: I1204 06:55:44.736137 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" podStartSLOduration=2.553870422 podStartE2EDuration="2.736116574s" podCreationTimestamp="2025-12-04 06:55:42 +0000 UTC" firstStartedPulling="2025-12-04 06:55:43.748079991 +0000 UTC m=+2680.936310766" lastFinishedPulling="2025-12-04 06:55:43.930326143 +0000 UTC m=+2681.118556918" observedRunningTime="2025-12-04 06:55:44.733400159 +0000 UTC m=+2681.921630944" watchObservedRunningTime="2025-12-04 06:55:44.736116574 +0000 UTC m=+2681.924347369" Dec 04 06:55:46 crc kubenswrapper[4685]: I1204 06:55:46.884948 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2dj9d"] Dec 04 06:55:46 crc kubenswrapper[4685]: I1204 06:55:46.887659 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:46 crc kubenswrapper[4685]: I1204 06:55:46.910448 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dj9d"] Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.064686 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-catalog-content\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.064791 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-utilities\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.064967 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khv6v\" (UniqueName: \"kubernetes.io/projected/42796d55-3e21-47d5-96fe-3bd6f52a019d-kube-api-access-khv6v\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.166231 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-catalog-content\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.166292 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-utilities\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.166369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khv6v\" (UniqueName: \"kubernetes.io/projected/42796d55-3e21-47d5-96fe-3bd6f52a019d-kube-api-access-khv6v\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.167039 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-catalog-content\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.167241 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-utilities\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.184294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khv6v\" (UniqueName: \"kubernetes.io/projected/42796d55-3e21-47d5-96fe-3bd6f52a019d-kube-api-access-khv6v\") pod \"redhat-marketplace-2dj9d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.209401 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:47 crc kubenswrapper[4685]: I1204 06:55:47.706705 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dj9d"] Dec 04 06:55:47 crc kubenswrapper[4685]: W1204 06:55:47.713675 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42796d55_3e21_47d5_96fe_3bd6f52a019d.slice/crio-67b39a5c8553df7abb3371ecf5712b9a9f7736ef28412df0ce5a90b8b335a161 WatchSource:0}: Error finding container 67b39a5c8553df7abb3371ecf5712b9a9f7736ef28412df0ce5a90b8b335a161: Status 404 returned error can't find the container with id 67b39a5c8553df7abb3371ecf5712b9a9f7736ef28412df0ce5a90b8b335a161 Dec 04 06:55:48 crc kubenswrapper[4685]: I1204 06:55:47.802734 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dj9d" event={"ID":"42796d55-3e21-47d5-96fe-3bd6f52a019d","Type":"ContainerStarted","Data":"67b39a5c8553df7abb3371ecf5712b9a9f7736ef28412df0ce5a90b8b335a161"} Dec 04 06:55:48 crc kubenswrapper[4685]: I1204 06:55:48.815194 4685 generic.go:334] "Generic (PLEG): container finished" podID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerID="b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2" exitCode=0 Dec 04 06:55:48 crc kubenswrapper[4685]: I1204 06:55:48.815270 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dj9d" event={"ID":"42796d55-3e21-47d5-96fe-3bd6f52a019d","Type":"ContainerDied","Data":"b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2"} Dec 04 06:55:49 crc kubenswrapper[4685]: I1204 06:55:49.828398 4685 generic.go:334] "Generic (PLEG): container finished" podID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerID="bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0" exitCode=0 Dec 04 06:55:49 crc kubenswrapper[4685]: I1204 06:55:49.828577 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dj9d" event={"ID":"42796d55-3e21-47d5-96fe-3bd6f52a019d","Type":"ContainerDied","Data":"bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0"} Dec 04 06:55:50 crc kubenswrapper[4685]: I1204 06:55:50.840037 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dj9d" event={"ID":"42796d55-3e21-47d5-96fe-3bd6f52a019d","Type":"ContainerStarted","Data":"6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2"} Dec 04 06:55:57 crc kubenswrapper[4685]: I1204 06:55:57.210035 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:57 crc kubenswrapper[4685]: I1204 06:55:57.210706 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:57 crc kubenswrapper[4685]: I1204 06:55:57.285016 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:57 crc kubenswrapper[4685]: I1204 06:55:57.301977 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2dj9d" podStartSLOduration=9.880864469 podStartE2EDuration="11.30195919s" podCreationTimestamp="2025-12-04 06:55:46 +0000 UTC" firstStartedPulling="2025-12-04 06:55:48.817138461 +0000 UTC m=+2686.005369236" lastFinishedPulling="2025-12-04 06:55:50.238233172 +0000 UTC m=+2687.426463957" observedRunningTime="2025-12-04 06:55:50.861961095 +0000 UTC m=+2688.050191870" watchObservedRunningTime="2025-12-04 06:55:57.30195919 +0000 UTC m=+2694.490189965" Dec 04 06:55:57 crc kubenswrapper[4685]: I1204 06:55:57.975545 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:55:58 crc kubenswrapper[4685]: I1204 06:55:58.042068 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dj9d"] Dec 04 06:55:59 crc kubenswrapper[4685]: I1204 06:55:59.923187 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2dj9d" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="registry-server" containerID="cri-o://6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2" gracePeriod=2 Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.375158 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.544175 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-catalog-content\") pod \"42796d55-3e21-47d5-96fe-3bd6f52a019d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.544346 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khv6v\" (UniqueName: \"kubernetes.io/projected/42796d55-3e21-47d5-96fe-3bd6f52a019d-kube-api-access-khv6v\") pod \"42796d55-3e21-47d5-96fe-3bd6f52a019d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.544443 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-utilities\") pod \"42796d55-3e21-47d5-96fe-3bd6f52a019d\" (UID: \"42796d55-3e21-47d5-96fe-3bd6f52a019d\") " Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.545416 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-utilities" (OuterVolumeSpecName: "utilities") pod "42796d55-3e21-47d5-96fe-3bd6f52a019d" (UID: "42796d55-3e21-47d5-96fe-3bd6f52a019d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.550900 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42796d55-3e21-47d5-96fe-3bd6f52a019d-kube-api-access-khv6v" (OuterVolumeSpecName: "kube-api-access-khv6v") pod "42796d55-3e21-47d5-96fe-3bd6f52a019d" (UID: "42796d55-3e21-47d5-96fe-3bd6f52a019d"). InnerVolumeSpecName "kube-api-access-khv6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.571085 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42796d55-3e21-47d5-96fe-3bd6f52a019d" (UID: "42796d55-3e21-47d5-96fe-3bd6f52a019d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.648360 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khv6v\" (UniqueName: \"kubernetes.io/projected/42796d55-3e21-47d5-96fe-3bd6f52a019d-kube-api-access-khv6v\") on node \"crc\" DevicePath \"\"" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.648527 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.648586 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42796d55-3e21-47d5-96fe-3bd6f52a019d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.936927 4685 generic.go:334] "Generic (PLEG): container finished" podID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerID="6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2" exitCode=0 Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.936972 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dj9d" event={"ID":"42796d55-3e21-47d5-96fe-3bd6f52a019d","Type":"ContainerDied","Data":"6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2"} Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.936998 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dj9d" event={"ID":"42796d55-3e21-47d5-96fe-3bd6f52a019d","Type":"ContainerDied","Data":"67b39a5c8553df7abb3371ecf5712b9a9f7736ef28412df0ce5a90b8b335a161"} Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.937016 4685 scope.go:117] "RemoveContainer" containerID="6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.937050 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dj9d" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.977583 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dj9d"] Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.985323 4685 scope.go:117] "RemoveContainer" containerID="bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0" Dec 04 06:56:00 crc kubenswrapper[4685]: I1204 06:56:00.985720 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dj9d"] Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.022938 4685 scope.go:117] "RemoveContainer" containerID="b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.057841 4685 scope.go:117] "RemoveContainer" containerID="6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2" Dec 04 06:56:01 crc kubenswrapper[4685]: E1204 06:56:01.058902 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2\": container with ID starting with 6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2 not found: ID does not exist" containerID="6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.058955 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2"} err="failed to get container status \"6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2\": rpc error: code = NotFound desc = could not find container \"6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2\": container with ID starting with 6ad48f949fa49271a4b47874ed2f113a5145e1dbc5fb3232e144cc19edf33ff2 not found: ID does not exist" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.058986 4685 scope.go:117] "RemoveContainer" containerID="bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0" Dec 04 06:56:01 crc kubenswrapper[4685]: E1204 06:56:01.059347 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0\": container with ID starting with bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0 not found: ID does not exist" containerID="bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.059380 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0"} err="failed to get container status \"bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0\": rpc error: code = NotFound desc = could not find container \"bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0\": container with ID starting with bc5306b16ed8abd1ddb8a4b36c8d4ce7e1bb22d8d08f7c971f80d27e7cd84cf0 not found: ID does not exist" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.059439 4685 scope.go:117] "RemoveContainer" containerID="b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2" Dec 04 06:56:01 crc kubenswrapper[4685]: E1204 06:56:01.059806 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2\": container with ID starting with b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2 not found: ID does not exist" containerID="b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.059844 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2"} err="failed to get container status \"b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2\": rpc error: code = NotFound desc = could not find container \"b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2\": container with ID starting with b55a5717570f37d7026c482b7d3d7c431ce840c8adae5c498ce32aa9e6277ce2 not found: ID does not exist" Dec 04 06:56:01 crc kubenswrapper[4685]: I1204 06:56:01.135968 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" path="/var/lib/kubelet/pods/42796d55-3e21-47d5-96fe-3bd6f52a019d/volumes" Dec 04 06:56:06 crc kubenswrapper[4685]: I1204 06:56:06.121212 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:56:06 crc kubenswrapper[4685]: I1204 06:56:06.121929 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.121114 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.121753 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.121818 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.122893 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"494f3cc360254283c3eea2f0c461cdde3ec1104223dd2529efff313cf274bc96"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.122977 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://494f3cc360254283c3eea2f0c461cdde3ec1104223dd2529efff313cf274bc96" gracePeriod=600 Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.282935 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="494f3cc360254283c3eea2f0c461cdde3ec1104223dd2529efff313cf274bc96" exitCode=0 Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.283074 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"494f3cc360254283c3eea2f0c461cdde3ec1104223dd2529efff313cf274bc96"} Dec 04 06:56:36 crc kubenswrapper[4685]: I1204 06:56:36.283627 4685 scope.go:117] "RemoveContainer" containerID="de92c84bfb99038f88ab85aceb8b3f18717d9bb173099858a8d22d694589e3f8" Dec 04 06:56:37 crc kubenswrapper[4685]: I1204 06:56:37.297524 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64"} Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.488866 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s2j8p"] Dec 04 06:57:15 crc kubenswrapper[4685]: E1204 06:57:15.490185 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="extract-utilities" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.490208 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="extract-utilities" Dec 04 06:57:15 crc kubenswrapper[4685]: E1204 06:57:15.490253 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="extract-content" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.490264 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="extract-content" Dec 04 06:57:15 crc kubenswrapper[4685]: E1204 06:57:15.490305 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="registry-server" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.490318 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="registry-server" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.490626 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="42796d55-3e21-47d5-96fe-3bd6f52a019d" containerName="registry-server" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.496917 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.514243 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2j8p"] Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.587336 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-catalog-content\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.587443 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvf7q\" (UniqueName: \"kubernetes.io/projected/91d55c52-dcec-40f3-b5f0-c6fff329cef1-kube-api-access-hvf7q\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.587552 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-utilities\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.689257 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvf7q\" (UniqueName: \"kubernetes.io/projected/91d55c52-dcec-40f3-b5f0-c6fff329cef1-kube-api-access-hvf7q\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.689361 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-utilities\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.689508 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-catalog-content\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.689930 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-utilities\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.689938 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-catalog-content\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.706898 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvf7q\" (UniqueName: \"kubernetes.io/projected/91d55c52-dcec-40f3-b5f0-c6fff329cef1-kube-api-access-hvf7q\") pod \"community-operators-s2j8p\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:15 crc kubenswrapper[4685]: I1204 06:57:15.860761 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:16 crc kubenswrapper[4685]: I1204 06:57:16.371943 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2j8p"] Dec 04 06:57:16 crc kubenswrapper[4685]: I1204 06:57:16.742698 4685 generic.go:334] "Generic (PLEG): container finished" podID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerID="4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e" exitCode=0 Dec 04 06:57:16 crc kubenswrapper[4685]: I1204 06:57:16.742953 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2j8p" event={"ID":"91d55c52-dcec-40f3-b5f0-c6fff329cef1","Type":"ContainerDied","Data":"4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e"} Dec 04 06:57:16 crc kubenswrapper[4685]: I1204 06:57:16.742980 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2j8p" event={"ID":"91d55c52-dcec-40f3-b5f0-c6fff329cef1","Type":"ContainerStarted","Data":"80a0b6e745bfd6baf29244e2ace6ece2f43dd3d5a308b293bdf7716a91693096"} Dec 04 06:57:17 crc kubenswrapper[4685]: I1204 06:57:17.755828 4685 generic.go:334] "Generic (PLEG): container finished" podID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerID="130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc" exitCode=0 Dec 04 06:57:17 crc kubenswrapper[4685]: I1204 06:57:17.755933 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2j8p" event={"ID":"91d55c52-dcec-40f3-b5f0-c6fff329cef1","Type":"ContainerDied","Data":"130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc"} Dec 04 06:57:18 crc kubenswrapper[4685]: I1204 06:57:18.767213 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2j8p" event={"ID":"91d55c52-dcec-40f3-b5f0-c6fff329cef1","Type":"ContainerStarted","Data":"a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63"} Dec 04 06:57:18 crc kubenswrapper[4685]: I1204 06:57:18.785107 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s2j8p" podStartSLOduration=2.315886918 podStartE2EDuration="3.785086858s" podCreationTimestamp="2025-12-04 06:57:15 +0000 UTC" firstStartedPulling="2025-12-04 06:57:16.745094806 +0000 UTC m=+2773.933325581" lastFinishedPulling="2025-12-04 06:57:18.214294716 +0000 UTC m=+2775.402525521" observedRunningTime="2025-12-04 06:57:18.782719494 +0000 UTC m=+2775.970950309" watchObservedRunningTime="2025-12-04 06:57:18.785086858 +0000 UTC m=+2775.973317633" Dec 04 06:57:25 crc kubenswrapper[4685]: I1204 06:57:25.860862 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:25 crc kubenswrapper[4685]: I1204 06:57:25.861575 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:25 crc kubenswrapper[4685]: I1204 06:57:25.928212 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:26 crc kubenswrapper[4685]: I1204 06:57:26.923834 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:26 crc kubenswrapper[4685]: I1204 06:57:26.981107 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s2j8p"] Dec 04 06:57:28 crc kubenswrapper[4685]: I1204 06:57:28.964796 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s2j8p" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="registry-server" containerID="cri-o://a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63" gracePeriod=2 Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.402823 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.564667 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvf7q\" (UniqueName: \"kubernetes.io/projected/91d55c52-dcec-40f3-b5f0-c6fff329cef1-kube-api-access-hvf7q\") pod \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.564729 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-catalog-content\") pod \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.564747 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-utilities\") pod \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\" (UID: \"91d55c52-dcec-40f3-b5f0-c6fff329cef1\") " Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.566152 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-utilities" (OuterVolumeSpecName: "utilities") pod "91d55c52-dcec-40f3-b5f0-c6fff329cef1" (UID: "91d55c52-dcec-40f3-b5f0-c6fff329cef1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.570527 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91d55c52-dcec-40f3-b5f0-c6fff329cef1-kube-api-access-hvf7q" (OuterVolumeSpecName: "kube-api-access-hvf7q") pod "91d55c52-dcec-40f3-b5f0-c6fff329cef1" (UID: "91d55c52-dcec-40f3-b5f0-c6fff329cef1"). InnerVolumeSpecName "kube-api-access-hvf7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.631458 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91d55c52-dcec-40f3-b5f0-c6fff329cef1" (UID: "91d55c52-dcec-40f3-b5f0-c6fff329cef1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.667307 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvf7q\" (UniqueName: \"kubernetes.io/projected/91d55c52-dcec-40f3-b5f0-c6fff329cef1-kube-api-access-hvf7q\") on node \"crc\" DevicePath \"\"" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.667338 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.667348 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d55c52-dcec-40f3-b5f0-c6fff329cef1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.976498 4685 generic.go:334] "Generic (PLEG): container finished" podID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerID="a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63" exitCode=0 Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.976537 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2j8p" event={"ID":"91d55c52-dcec-40f3-b5f0-c6fff329cef1","Type":"ContainerDied","Data":"a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63"} Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.976561 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2j8p" event={"ID":"91d55c52-dcec-40f3-b5f0-c6fff329cef1","Type":"ContainerDied","Data":"80a0b6e745bfd6baf29244e2ace6ece2f43dd3d5a308b293bdf7716a91693096"} Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.976577 4685 scope.go:117] "RemoveContainer" containerID="a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63" Dec 04 06:57:29 crc kubenswrapper[4685]: I1204 06:57:29.976677 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2j8p" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.006655 4685 scope.go:117] "RemoveContainer" containerID="130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.014347 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s2j8p"] Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.028715 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s2j8p"] Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.045455 4685 scope.go:117] "RemoveContainer" containerID="4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.088709 4685 scope.go:117] "RemoveContainer" containerID="a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63" Dec 04 06:57:30 crc kubenswrapper[4685]: E1204 06:57:30.089282 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63\": container with ID starting with a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63 not found: ID does not exist" containerID="a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.089318 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63"} err="failed to get container status \"a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63\": rpc error: code = NotFound desc = could not find container \"a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63\": container with ID starting with a02c2216f543c297e1c6cd11b2ad3deed20d1b989ceab5e0395a20a701bb5e63 not found: ID does not exist" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.089344 4685 scope.go:117] "RemoveContainer" containerID="130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc" Dec 04 06:57:30 crc kubenswrapper[4685]: E1204 06:57:30.089826 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc\": container with ID starting with 130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc not found: ID does not exist" containerID="130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.089853 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc"} err="failed to get container status \"130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc\": rpc error: code = NotFound desc = could not find container \"130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc\": container with ID starting with 130ac200d8ae327601e0456edd2b8ebce967199430f3edc26bfe0a7a6d0e7ebc not found: ID does not exist" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.089873 4685 scope.go:117] "RemoveContainer" containerID="4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e" Dec 04 06:57:30 crc kubenswrapper[4685]: E1204 06:57:30.090184 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e\": container with ID starting with 4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e not found: ID does not exist" containerID="4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e" Dec 04 06:57:30 crc kubenswrapper[4685]: I1204 06:57:30.090210 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e"} err="failed to get container status \"4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e\": rpc error: code = NotFound desc = could not find container \"4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e\": container with ID starting with 4e7b3f71267ee2ee5bebf2915023531c7bebb24d1d499d1c60c762fa732ec43e not found: ID does not exist" Dec 04 06:57:31 crc kubenswrapper[4685]: I1204 06:57:31.141481 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" path="/var/lib/kubelet/pods/91d55c52-dcec-40f3-b5f0-c6fff329cef1/volumes" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.328205 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-np4k4"] Dec 04 06:57:46 crc kubenswrapper[4685]: E1204 06:57:46.329790 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="registry-server" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.329808 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="registry-server" Dec 04 06:57:46 crc kubenswrapper[4685]: E1204 06:57:46.329848 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="extract-utilities" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.329856 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="extract-utilities" Dec 04 06:57:46 crc kubenswrapper[4685]: E1204 06:57:46.329867 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="extract-content" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.329875 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="extract-content" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.330113 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d55c52-dcec-40f3-b5f0-c6fff329cef1" containerName="registry-server" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.332232 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.360315 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-np4k4"] Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.380112 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-utilities\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.380294 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s26zv\" (UniqueName: \"kubernetes.io/projected/1b3217ce-70a5-4a17-aa55-54bbbc13384f-kube-api-access-s26zv\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.380448 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-catalog-content\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.482452 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-catalog-content\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.482546 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-utilities\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.483333 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s26zv\" (UniqueName: \"kubernetes.io/projected/1b3217ce-70a5-4a17-aa55-54bbbc13384f-kube-api-access-s26zv\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.484245 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-utilities\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.484452 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-catalog-content\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.505448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s26zv\" (UniqueName: \"kubernetes.io/projected/1b3217ce-70a5-4a17-aa55-54bbbc13384f-kube-api-access-s26zv\") pod \"redhat-operators-np4k4\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:46 crc kubenswrapper[4685]: I1204 06:57:46.714505 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.209150 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-np4k4"] Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.318082 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9jjx4"] Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.322195 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.331087 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9jjx4"] Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.389572 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerStarted","Data":"55e1c32e0c257b4d899fce084cbf28bfe53dd101fa54e9cde680c60f0dcb57f4"} Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.406453 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-utilities\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.406561 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxb89\" (UniqueName: \"kubernetes.io/projected/0fece3d4-42c6-4f05-9046-ed317ced462a-kube-api-access-wxb89\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.406653 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-catalog-content\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.508225 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-catalog-content\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.508304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-utilities\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.508370 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxb89\" (UniqueName: \"kubernetes.io/projected/0fece3d4-42c6-4f05-9046-ed317ced462a-kube-api-access-wxb89\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.508844 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-catalog-content\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.508949 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-utilities\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.535986 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxb89\" (UniqueName: \"kubernetes.io/projected/0fece3d4-42c6-4f05-9046-ed317ced462a-kube-api-access-wxb89\") pod \"certified-operators-9jjx4\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:47 crc kubenswrapper[4685]: I1204 06:57:47.662079 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:48 crc kubenswrapper[4685]: I1204 06:57:48.155495 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9jjx4"] Dec 04 06:57:48 crc kubenswrapper[4685]: I1204 06:57:48.399744 4685 generic.go:334] "Generic (PLEG): container finished" podID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerID="6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2" exitCode=0 Dec 04 06:57:48 crc kubenswrapper[4685]: I1204 06:57:48.399926 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerDied","Data":"6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2"} Dec 04 06:57:48 crc kubenswrapper[4685]: I1204 06:57:48.402442 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerStarted","Data":"4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02"} Dec 04 06:57:48 crc kubenswrapper[4685]: I1204 06:57:48.402518 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerStarted","Data":"ba68dc23e2395742dbdbb251ace7569df2185153a4d3b7c1378c6af33d996559"} Dec 04 06:57:48 crc kubenswrapper[4685]: I1204 06:57:48.402621 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 06:57:49 crc kubenswrapper[4685]: I1204 06:57:49.412217 4685 generic.go:334] "Generic (PLEG): container finished" podID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerID="4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02" exitCode=0 Dec 04 06:57:49 crc kubenswrapper[4685]: I1204 06:57:49.412310 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerDied","Data":"4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02"} Dec 04 06:57:49 crc kubenswrapper[4685]: I1204 06:57:49.414977 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerStarted","Data":"a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4"} Dec 04 06:57:51 crc kubenswrapper[4685]: I1204 06:57:51.444746 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerStarted","Data":"89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7"} Dec 04 06:57:51 crc kubenswrapper[4685]: I1204 06:57:51.449097 4685 generic.go:334] "Generic (PLEG): container finished" podID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerID="a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4" exitCode=0 Dec 04 06:57:51 crc kubenswrapper[4685]: I1204 06:57:51.449149 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerDied","Data":"a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4"} Dec 04 06:57:52 crc kubenswrapper[4685]: I1204 06:57:52.463356 4685 generic.go:334] "Generic (PLEG): container finished" podID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerID="89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7" exitCode=0 Dec 04 06:57:52 crc kubenswrapper[4685]: I1204 06:57:52.463437 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerDied","Data":"89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7"} Dec 04 06:57:53 crc kubenswrapper[4685]: I1204 06:57:53.475294 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerStarted","Data":"42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c"} Dec 04 06:57:53 crc kubenswrapper[4685]: I1204 06:57:53.479121 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerStarted","Data":"43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161"} Dec 04 06:57:53 crc kubenswrapper[4685]: I1204 06:57:53.508560 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9jjx4" podStartSLOduration=3.024938129 podStartE2EDuration="6.508524018s" podCreationTimestamp="2025-12-04 06:57:47 +0000 UTC" firstStartedPulling="2025-12-04 06:57:49.414223444 +0000 UTC m=+2806.602454209" lastFinishedPulling="2025-12-04 06:57:52.897809323 +0000 UTC m=+2810.086040098" observedRunningTime="2025-12-04 06:57:53.4964825 +0000 UTC m=+2810.684713295" watchObservedRunningTime="2025-12-04 06:57:53.508524018 +0000 UTC m=+2810.696754823" Dec 04 06:57:53 crc kubenswrapper[4685]: I1204 06:57:53.521801 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-np4k4" podStartSLOduration=3.940125626 podStartE2EDuration="7.521781894s" podCreationTimestamp="2025-12-04 06:57:46 +0000 UTC" firstStartedPulling="2025-12-04 06:57:48.402292991 +0000 UTC m=+2805.590523766" lastFinishedPulling="2025-12-04 06:57:51.983949259 +0000 UTC m=+2809.172180034" observedRunningTime="2025-12-04 06:57:53.515787616 +0000 UTC m=+2810.704018421" watchObservedRunningTime="2025-12-04 06:57:53.521781894 +0000 UTC m=+2810.710012669" Dec 04 06:57:56 crc kubenswrapper[4685]: I1204 06:57:56.714881 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:56 crc kubenswrapper[4685]: I1204 06:57:56.715495 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:57:57 crc kubenswrapper[4685]: I1204 06:57:57.662715 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:57 crc kubenswrapper[4685]: I1204 06:57:57.662778 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:57 crc kubenswrapper[4685]: I1204 06:57:57.727823 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:57 crc kubenswrapper[4685]: I1204 06:57:57.761646 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-np4k4" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="registry-server" probeResult="failure" output=< Dec 04 06:57:57 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Dec 04 06:57:57 crc kubenswrapper[4685]: > Dec 04 06:57:58 crc kubenswrapper[4685]: I1204 06:57:58.608627 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:57:59 crc kubenswrapper[4685]: I1204 06:57:59.710979 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9jjx4"] Dec 04 06:58:00 crc kubenswrapper[4685]: I1204 06:58:00.553127 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9jjx4" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="registry-server" containerID="cri-o://42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c" gracePeriod=2 Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.064354 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.190671 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-catalog-content\") pod \"0fece3d4-42c6-4f05-9046-ed317ced462a\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.191220 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxb89\" (UniqueName: \"kubernetes.io/projected/0fece3d4-42c6-4f05-9046-ed317ced462a-kube-api-access-wxb89\") pod \"0fece3d4-42c6-4f05-9046-ed317ced462a\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.191268 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-utilities\") pod \"0fece3d4-42c6-4f05-9046-ed317ced462a\" (UID: \"0fece3d4-42c6-4f05-9046-ed317ced462a\") " Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.192122 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-utilities" (OuterVolumeSpecName: "utilities") pod "0fece3d4-42c6-4f05-9046-ed317ced462a" (UID: "0fece3d4-42c6-4f05-9046-ed317ced462a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.196476 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fece3d4-42c6-4f05-9046-ed317ced462a-kube-api-access-wxb89" (OuterVolumeSpecName: "kube-api-access-wxb89") pod "0fece3d4-42c6-4f05-9046-ed317ced462a" (UID: "0fece3d4-42c6-4f05-9046-ed317ced462a"). InnerVolumeSpecName "kube-api-access-wxb89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.234458 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fece3d4-42c6-4f05-9046-ed317ced462a" (UID: "0fece3d4-42c6-4f05-9046-ed317ced462a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.293964 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.294003 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxb89\" (UniqueName: \"kubernetes.io/projected/0fece3d4-42c6-4f05-9046-ed317ced462a-kube-api-access-wxb89\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.294015 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fece3d4-42c6-4f05-9046-ed317ced462a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.571098 4685 generic.go:334] "Generic (PLEG): container finished" podID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerID="42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c" exitCode=0 Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.571150 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerDied","Data":"42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c"} Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.571550 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jjx4" event={"ID":"0fece3d4-42c6-4f05-9046-ed317ced462a","Type":"ContainerDied","Data":"ba68dc23e2395742dbdbb251ace7569df2185153a4d3b7c1378c6af33d996559"} Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.571581 4685 scope.go:117] "RemoveContainer" containerID="42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.571197 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jjx4" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.610369 4685 scope.go:117] "RemoveContainer" containerID="89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.622721 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9jjx4"] Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.635707 4685 scope.go:117] "RemoveContainer" containerID="4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.636071 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9jjx4"] Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.698950 4685 scope.go:117] "RemoveContainer" containerID="42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c" Dec 04 06:58:01 crc kubenswrapper[4685]: E1204 06:58:01.699513 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c\": container with ID starting with 42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c not found: ID does not exist" containerID="42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.699559 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c"} err="failed to get container status \"42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c\": rpc error: code = NotFound desc = could not find container \"42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c\": container with ID starting with 42c56ecb3d8ea18c4158e43b52eb53930d1b755ffad5339b425513ad43c03e3c not found: ID does not exist" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.699586 4685 scope.go:117] "RemoveContainer" containerID="89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7" Dec 04 06:58:01 crc kubenswrapper[4685]: E1204 06:58:01.700068 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7\": container with ID starting with 89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7 not found: ID does not exist" containerID="89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.700132 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7"} err="failed to get container status \"89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7\": rpc error: code = NotFound desc = could not find container \"89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7\": container with ID starting with 89e01f2d89cd557affb3fea75442cae028ac9d87dbbd4b361dafe5d4f526b8b7 not found: ID does not exist" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.700189 4685 scope.go:117] "RemoveContainer" containerID="4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02" Dec 04 06:58:01 crc kubenswrapper[4685]: E1204 06:58:01.700655 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02\": container with ID starting with 4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02 not found: ID does not exist" containerID="4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02" Dec 04 06:58:01 crc kubenswrapper[4685]: I1204 06:58:01.700705 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02"} err="failed to get container status \"4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02\": rpc error: code = NotFound desc = could not find container \"4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02\": container with ID starting with 4688e70cd04732d76c9e052213aa75433aaeb3b77c335481b5efdc60dc1c3f02 not found: ID does not exist" Dec 04 06:58:03 crc kubenswrapper[4685]: I1204 06:58:03.147713 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" path="/var/lib/kubelet/pods/0fece3d4-42c6-4f05-9046-ed317ced462a/volumes" Dec 04 06:58:06 crc kubenswrapper[4685]: I1204 06:58:06.794384 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:58:06 crc kubenswrapper[4685]: I1204 06:58:06.868487 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:58:07 crc kubenswrapper[4685]: I1204 06:58:07.047007 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-np4k4"] Dec 04 06:58:08 crc kubenswrapper[4685]: I1204 06:58:08.666504 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-np4k4" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="registry-server" containerID="cri-o://43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161" gracePeriod=2 Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.152476 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.265106 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-utilities\") pod \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.265220 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-catalog-content\") pod \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.265345 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s26zv\" (UniqueName: \"kubernetes.io/projected/1b3217ce-70a5-4a17-aa55-54bbbc13384f-kube-api-access-s26zv\") pod \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\" (UID: \"1b3217ce-70a5-4a17-aa55-54bbbc13384f\") " Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.266932 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-utilities" (OuterVolumeSpecName: "utilities") pod "1b3217ce-70a5-4a17-aa55-54bbbc13384f" (UID: "1b3217ce-70a5-4a17-aa55-54bbbc13384f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.273090 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b3217ce-70a5-4a17-aa55-54bbbc13384f-kube-api-access-s26zv" (OuterVolumeSpecName: "kube-api-access-s26zv") pod "1b3217ce-70a5-4a17-aa55-54bbbc13384f" (UID: "1b3217ce-70a5-4a17-aa55-54bbbc13384f"). InnerVolumeSpecName "kube-api-access-s26zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.368884 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s26zv\" (UniqueName: \"kubernetes.io/projected/1b3217ce-70a5-4a17-aa55-54bbbc13384f-kube-api-access-s26zv\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.368918 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.371547 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b3217ce-70a5-4a17-aa55-54bbbc13384f" (UID: "1b3217ce-70a5-4a17-aa55-54bbbc13384f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.471331 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3217ce-70a5-4a17-aa55-54bbbc13384f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.678252 4685 generic.go:334] "Generic (PLEG): container finished" podID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerID="43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161" exitCode=0 Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.678303 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerDied","Data":"43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161"} Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.678365 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-np4k4" event={"ID":"1b3217ce-70a5-4a17-aa55-54bbbc13384f","Type":"ContainerDied","Data":"55e1c32e0c257b4d899fce084cbf28bfe53dd101fa54e9cde680c60f0dcb57f4"} Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.678393 4685 scope.go:117] "RemoveContainer" containerID="43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.680630 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-np4k4" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.695803 4685 scope.go:117] "RemoveContainer" containerID="a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.729852 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-np4k4"] Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.737123 4685 scope.go:117] "RemoveContainer" containerID="6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.740297 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-np4k4"] Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.779461 4685 scope.go:117] "RemoveContainer" containerID="43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161" Dec 04 06:58:09 crc kubenswrapper[4685]: E1204 06:58:09.779883 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161\": container with ID starting with 43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161 not found: ID does not exist" containerID="43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.779912 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161"} err="failed to get container status \"43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161\": rpc error: code = NotFound desc = could not find container \"43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161\": container with ID starting with 43b6848524a5dbdfaff5683efb58694a248ba645cf7296309dd8ba3a336c7161 not found: ID does not exist" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.779934 4685 scope.go:117] "RemoveContainer" containerID="a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4" Dec 04 06:58:09 crc kubenswrapper[4685]: E1204 06:58:09.780422 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4\": container with ID starting with a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4 not found: ID does not exist" containerID="a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.780450 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4"} err="failed to get container status \"a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4\": rpc error: code = NotFound desc = could not find container \"a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4\": container with ID starting with a806da777db42424e792aca2b70394f536ab7ef6e5b84eddcdf0f2fdc4ecb9c4 not found: ID does not exist" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.780467 4685 scope.go:117] "RemoveContainer" containerID="6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2" Dec 04 06:58:09 crc kubenswrapper[4685]: E1204 06:58:09.780712 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2\": container with ID starting with 6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2 not found: ID does not exist" containerID="6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2" Dec 04 06:58:09 crc kubenswrapper[4685]: I1204 06:58:09.780787 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2"} err="failed to get container status \"6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2\": rpc error: code = NotFound desc = could not find container \"6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2\": container with ID starting with 6527684ec8901c4d9fa2ca718e483ae8102bedb197dd2fd859660a1d8589ddb2 not found: ID does not exist" Dec 04 06:58:11 crc kubenswrapper[4685]: I1204 06:58:11.144405 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" path="/var/lib/kubelet/pods/1b3217ce-70a5-4a17-aa55-54bbbc13384f/volumes" Dec 04 06:58:36 crc kubenswrapper[4685]: I1204 06:58:36.121784 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:58:36 crc kubenswrapper[4685]: I1204 06:58:36.122320 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:58:42 crc kubenswrapper[4685]: I1204 06:58:42.997093 4685 generic.go:334] "Generic (PLEG): container finished" podID="175a4186-effa-431c-9808-33f3fa73119b" containerID="8a8f73b62b67382a4801a0a47fbbbe4081155a86a43488323a9c651036f0cbad" exitCode=0 Dec 04 06:58:42 crc kubenswrapper[4685]: I1204 06:58:42.997727 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" event={"ID":"175a4186-effa-431c-9808-33f3fa73119b","Type":"ContainerDied","Data":"8a8f73b62b67382a4801a0a47fbbbe4081155a86a43488323a9c651036f0cbad"} Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.433450 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.511077 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-inventory\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.511170 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-telemetry-combined-ca-bundle\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.511200 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-2\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.511999 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/175a4186-effa-431c-9808-33f3fa73119b-kube-api-access-d58r4\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.512092 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-0\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.512125 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-1\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.512192 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ssh-key\") pod \"175a4186-effa-431c-9808-33f3fa73119b\" (UID: \"175a4186-effa-431c-9808-33f3fa73119b\") " Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.516936 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175a4186-effa-431c-9808-33f3fa73119b-kube-api-access-d58r4" (OuterVolumeSpecName: "kube-api-access-d58r4") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "kube-api-access-d58r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.517354 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.544978 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.545365 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-inventory" (OuterVolumeSpecName: "inventory") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.545380 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.547262 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.552087 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "175a4186-effa-431c-9808-33f3fa73119b" (UID: "175a4186-effa-431c-9808-33f3fa73119b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613819 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/175a4186-effa-431c-9808-33f3fa73119b-kube-api-access-d58r4\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613847 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613857 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613867 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613877 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613886 4685 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:44 crc kubenswrapper[4685]: I1204 06:58:44.613897 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/175a4186-effa-431c-9808-33f3fa73119b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 06:58:45 crc kubenswrapper[4685]: I1204 06:58:45.019031 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" event={"ID":"175a4186-effa-431c-9808-33f3fa73119b","Type":"ContainerDied","Data":"baef2a430f8266483c16cace0782ad5784176214c49fae97f1f33f8cc539c356"} Dec 04 06:58:45 crc kubenswrapper[4685]: I1204 06:58:45.019435 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baef2a430f8266483c16cace0782ad5784176214c49fae97f1f33f8cc539c356" Dec 04 06:58:45 crc kubenswrapper[4685]: I1204 06:58:45.019145 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4" Dec 04 06:59:06 crc kubenswrapper[4685]: I1204 06:59:06.121662 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:59:06 crc kubenswrapper[4685]: I1204 06:59:06.122228 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.899360 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900507 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="extract-utilities" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900528 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="extract-utilities" Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900542 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="registry-server" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900552 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="registry-server" Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900574 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="extract-content" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900583 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="extract-content" Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900592 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="registry-server" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900601 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="registry-server" Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900625 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="extract-utilities" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900633 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="extract-utilities" Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900659 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175a4186-effa-431c-9808-33f3fa73119b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900669 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="175a4186-effa-431c-9808-33f3fa73119b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 06:59:34 crc kubenswrapper[4685]: E1204 06:59:34.900678 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="extract-content" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900686 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="extract-content" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900893 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fece3d4-42c6-4f05-9046-ed317ced462a" containerName="registry-server" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900927 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3217ce-70a5-4a17-aa55-54bbbc13384f" containerName="registry-server" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.900959 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="175a4186-effa-431c-9808-33f3fa73119b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.901772 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.904138 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.904204 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.905132 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fdj2s" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.905207 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 04 06:59:34 crc kubenswrapper[4685]: I1204 06:59:34.909762 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079458 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079542 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079594 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079642 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079703 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd4b4\" (UniqueName: \"kubernetes.io/projected/c8fff46b-61e1-48da-a047-9ace06e11177-kube-api-access-pd4b4\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079729 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079752 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079812 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.079866 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-config-data\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.181870 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.181940 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.182816 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.182902 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.183251 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd4b4\" (UniqueName: \"kubernetes.io/projected/c8fff46b-61e1-48da-a047-9ace06e11177-kube-api-access-pd4b4\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.183655 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.183194 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.183683 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.183654 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.183960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.184094 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-config-data\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.184434 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.184794 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.185223 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-config-data\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.190563 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.191115 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.193030 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.202689 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd4b4\" (UniqueName: \"kubernetes.io/projected/c8fff46b-61e1-48da-a047-9ace06e11177-kube-api-access-pd4b4\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.223127 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " pod="openstack/tempest-tests-tempest" Dec 04 06:59:35 crc kubenswrapper[4685]: I1204 06:59:35.526228 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.002915 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.121426 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.121780 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.121824 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.122385 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.122481 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" gracePeriod=600 Dec 04 06:59:36 crc kubenswrapper[4685]: E1204 06:59:36.254740 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.579046 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c8fff46b-61e1-48da-a047-9ace06e11177","Type":"ContainerStarted","Data":"f639b34c380e5e7348937a7306792d23c7f8b8f56c678f6a5cb013e3affdeabc"} Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.583814 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" exitCode=0 Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.583844 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64"} Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.583919 4685 scope.go:117] "RemoveContainer" containerID="494f3cc360254283c3eea2f0c461cdde3ec1104223dd2529efff313cf274bc96" Dec 04 06:59:36 crc kubenswrapper[4685]: I1204 06:59:36.585052 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 06:59:36 crc kubenswrapper[4685]: E1204 06:59:36.585670 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 06:59:51 crc kubenswrapper[4685]: I1204 06:59:51.126133 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 06:59:51 crc kubenswrapper[4685]: E1204 06:59:51.127006 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.164129 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf"] Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.166592 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.171690 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.171867 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.173310 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf"] Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.294639 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29bc9fde-ad00-4993-a763-792d2c83e9d1-secret-volume\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.294862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ttsq\" (UniqueName: \"kubernetes.io/projected/29bc9fde-ad00-4993-a763-792d2c83e9d1-kube-api-access-2ttsq\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.294882 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29bc9fde-ad00-4993-a763-792d2c83e9d1-config-volume\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.396835 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29bc9fde-ad00-4993-a763-792d2c83e9d1-secret-volume\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.396986 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ttsq\" (UniqueName: \"kubernetes.io/projected/29bc9fde-ad00-4993-a763-792d2c83e9d1-kube-api-access-2ttsq\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.397011 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29bc9fde-ad00-4993-a763-792d2c83e9d1-config-volume\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.398009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29bc9fde-ad00-4993-a763-792d2c83e9d1-config-volume\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.403582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29bc9fde-ad00-4993-a763-792d2c83e9d1-secret-volume\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.414594 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ttsq\" (UniqueName: \"kubernetes.io/projected/29bc9fde-ad00-4993-a763-792d2c83e9d1-kube-api-access-2ttsq\") pod \"collect-profiles-29413860-cwxxf\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:00 crc kubenswrapper[4685]: I1204 07:00:00.500207 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:02 crc kubenswrapper[4685]: I1204 07:00:02.126054 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:00:02 crc kubenswrapper[4685]: E1204 07:00:02.126725 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:00:04 crc kubenswrapper[4685]: E1204 07:00:04.198050 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 04 07:00:04 crc kubenswrapper[4685]: E1204 07:00:04.198557 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pd4b4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(c8fff46b-61e1-48da-a047-9ace06e11177): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 07:00:04 crc kubenswrapper[4685]: E1204 07:00:04.199769 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="c8fff46b-61e1-48da-a047-9ace06e11177" Dec 04 07:00:04 crc kubenswrapper[4685]: E1204 07:00:04.237173 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="c8fff46b-61e1-48da-a047-9ace06e11177" Dec 04 07:00:04 crc kubenswrapper[4685]: I1204 07:00:04.643165 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf"] Dec 04 07:00:05 crc kubenswrapper[4685]: I1204 07:00:05.249835 4685 generic.go:334] "Generic (PLEG): container finished" podID="29bc9fde-ad00-4993-a763-792d2c83e9d1" containerID="7cd3b530d7b4f04535762e425f3dd0e98bf858d5642cbc386081706dfcba4a9c" exitCode=0 Dec 04 07:00:05 crc kubenswrapper[4685]: I1204 07:00:05.249931 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" event={"ID":"29bc9fde-ad00-4993-a763-792d2c83e9d1","Type":"ContainerDied","Data":"7cd3b530d7b4f04535762e425f3dd0e98bf858d5642cbc386081706dfcba4a9c"} Dec 04 07:00:05 crc kubenswrapper[4685]: I1204 07:00:05.250114 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" event={"ID":"29bc9fde-ad00-4993-a763-792d2c83e9d1","Type":"ContainerStarted","Data":"9702cf412f6fe7c7297403209c0070f68ddba84158833e84fb7228ddeb7a2a4b"} Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.665456 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.820560 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29bc9fde-ad00-4993-a763-792d2c83e9d1-secret-volume\") pod \"29bc9fde-ad00-4993-a763-792d2c83e9d1\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.820770 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29bc9fde-ad00-4993-a763-792d2c83e9d1-config-volume\") pod \"29bc9fde-ad00-4993-a763-792d2c83e9d1\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.820871 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ttsq\" (UniqueName: \"kubernetes.io/projected/29bc9fde-ad00-4993-a763-792d2c83e9d1-kube-api-access-2ttsq\") pod \"29bc9fde-ad00-4993-a763-792d2c83e9d1\" (UID: \"29bc9fde-ad00-4993-a763-792d2c83e9d1\") " Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.822494 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29bc9fde-ad00-4993-a763-792d2c83e9d1-config-volume" (OuterVolumeSpecName: "config-volume") pod "29bc9fde-ad00-4993-a763-792d2c83e9d1" (UID: "29bc9fde-ad00-4993-a763-792d2c83e9d1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.827908 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bc9fde-ad00-4993-a763-792d2c83e9d1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "29bc9fde-ad00-4993-a763-792d2c83e9d1" (UID: "29bc9fde-ad00-4993-a763-792d2c83e9d1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.828153 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29bc9fde-ad00-4993-a763-792d2c83e9d1-kube-api-access-2ttsq" (OuterVolumeSpecName: "kube-api-access-2ttsq") pod "29bc9fde-ad00-4993-a763-792d2c83e9d1" (UID: "29bc9fde-ad00-4993-a763-792d2c83e9d1"). InnerVolumeSpecName "kube-api-access-2ttsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.922841 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29bc9fde-ad00-4993-a763-792d2c83e9d1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.922871 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ttsq\" (UniqueName: \"kubernetes.io/projected/29bc9fde-ad00-4993-a763-792d2c83e9d1-kube-api-access-2ttsq\") on node \"crc\" DevicePath \"\"" Dec 04 07:00:06 crc kubenswrapper[4685]: I1204 07:00:06.922882 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29bc9fde-ad00-4993-a763-792d2c83e9d1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 07:00:07 crc kubenswrapper[4685]: I1204 07:00:07.274193 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" event={"ID":"29bc9fde-ad00-4993-a763-792d2c83e9d1","Type":"ContainerDied","Data":"9702cf412f6fe7c7297403209c0070f68ddba84158833e84fb7228ddeb7a2a4b"} Dec 04 07:00:07 crc kubenswrapper[4685]: I1204 07:00:07.274627 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9702cf412f6fe7c7297403209c0070f68ddba84158833e84fb7228ddeb7a2a4b" Dec 04 07:00:07 crc kubenswrapper[4685]: I1204 07:00:07.274237 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413860-cwxxf" Dec 04 07:00:07 crc kubenswrapper[4685]: I1204 07:00:07.747789 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss"] Dec 04 07:00:07 crc kubenswrapper[4685]: I1204 07:00:07.755713 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413815-cktss"] Dec 04 07:00:09 crc kubenswrapper[4685]: I1204 07:00:09.142329 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecaa27dd-6791-4c59-8c1e-7fae7c341d9a" path="/var/lib/kubelet/pods/ecaa27dd-6791-4c59-8c1e-7fae7c341d9a/volumes" Dec 04 07:00:15 crc kubenswrapper[4685]: I1204 07:00:15.125700 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:00:15 crc kubenswrapper[4685]: E1204 07:00:15.126493 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:00:18 crc kubenswrapper[4685]: I1204 07:00:18.625851 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 07:00:19 crc kubenswrapper[4685]: I1204 07:00:19.218441 4685 scope.go:117] "RemoveContainer" containerID="73cae50a1c794021b3b7f17a6cbfd5b48162d5014ceb0f37a2b7dd40e9ab03f8" Dec 04 07:00:20 crc kubenswrapper[4685]: I1204 07:00:20.402007 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c8fff46b-61e1-48da-a047-9ace06e11177","Type":"ContainerStarted","Data":"f9039625c021671b160c25e306848ed8932c307e304aef00bcb5172eb7ddff90"} Dec 04 07:00:20 crc kubenswrapper[4685]: I1204 07:00:20.424793 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.817352716 podStartE2EDuration="47.424775947s" podCreationTimestamp="2025-12-04 06:59:33 +0000 UTC" firstStartedPulling="2025-12-04 06:59:36.015304087 +0000 UTC m=+2913.203534862" lastFinishedPulling="2025-12-04 07:00:18.622727308 +0000 UTC m=+2955.810958093" observedRunningTime="2025-12-04 07:00:20.42359444 +0000 UTC m=+2957.611825225" watchObservedRunningTime="2025-12-04 07:00:20.424775947 +0000 UTC m=+2957.613006732" Dec 04 07:00:27 crc kubenswrapper[4685]: I1204 07:00:27.126785 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:00:27 crc kubenswrapper[4685]: E1204 07:00:27.127633 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:00:42 crc kubenswrapper[4685]: I1204 07:00:42.126050 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:00:42 crc kubenswrapper[4685]: E1204 07:00:42.126837 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:00:56 crc kubenswrapper[4685]: I1204 07:00:56.126519 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:00:56 crc kubenswrapper[4685]: E1204 07:00:56.127429 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.150846 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413861-26zsj"] Dec 04 07:01:00 crc kubenswrapper[4685]: E1204 07:01:00.151911 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29bc9fde-ad00-4993-a763-792d2c83e9d1" containerName="collect-profiles" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.151926 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="29bc9fde-ad00-4993-a763-792d2c83e9d1" containerName="collect-profiles" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.152198 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="29bc9fde-ad00-4993-a763-792d2c83e9d1" containerName="collect-profiles" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.153093 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.165921 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413861-26zsj"] Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.228265 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-fernet-keys\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.228391 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-config-data\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.228436 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-combined-ca-bundle\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.228530 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv2br\" (UniqueName: \"kubernetes.io/projected/f45f0571-fa49-4f6a-a36d-82dadc904495-kube-api-access-lv2br\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.330666 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-fernet-keys\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.330792 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-config-data\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.330822 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-combined-ca-bundle\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.330865 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv2br\" (UniqueName: \"kubernetes.io/projected/f45f0571-fa49-4f6a-a36d-82dadc904495-kube-api-access-lv2br\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.336780 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-combined-ca-bundle\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.347250 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-fernet-keys\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.347703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv2br\" (UniqueName: \"kubernetes.io/projected/f45f0571-fa49-4f6a-a36d-82dadc904495-kube-api-access-lv2br\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.358093 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-config-data\") pod \"keystone-cron-29413861-26zsj\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.474565 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:00 crc kubenswrapper[4685]: I1204 07:01:00.982653 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413861-26zsj"] Dec 04 07:01:01 crc kubenswrapper[4685]: I1204 07:01:01.814736 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413861-26zsj" event={"ID":"f45f0571-fa49-4f6a-a36d-82dadc904495","Type":"ContainerStarted","Data":"2adec322531760635e196429b267f08a18bc2b93cafee2ca629fa1754861909a"} Dec 04 07:01:01 crc kubenswrapper[4685]: I1204 07:01:01.815052 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413861-26zsj" event={"ID":"f45f0571-fa49-4f6a-a36d-82dadc904495","Type":"ContainerStarted","Data":"4ff354e3134557f79822d53991223b0e87cc44c5c9962b63b325c7045025caaf"} Dec 04 07:01:01 crc kubenswrapper[4685]: I1204 07:01:01.852180 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29413861-26zsj" podStartSLOduration=1.852152172 podStartE2EDuration="1.852152172s" podCreationTimestamp="2025-12-04 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 07:01:01.828959984 +0000 UTC m=+2999.017190759" watchObservedRunningTime="2025-12-04 07:01:01.852152172 +0000 UTC m=+2999.040382947" Dec 04 07:01:03 crc kubenswrapper[4685]: I1204 07:01:03.838457 4685 generic.go:334] "Generic (PLEG): container finished" podID="f45f0571-fa49-4f6a-a36d-82dadc904495" containerID="2adec322531760635e196429b267f08a18bc2b93cafee2ca629fa1754861909a" exitCode=0 Dec 04 07:01:03 crc kubenswrapper[4685]: I1204 07:01:03.838573 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413861-26zsj" event={"ID":"f45f0571-fa49-4f6a-a36d-82dadc904495","Type":"ContainerDied","Data":"2adec322531760635e196429b267f08a18bc2b93cafee2ca629fa1754861909a"} Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.205553 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.374050 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv2br\" (UniqueName: \"kubernetes.io/projected/f45f0571-fa49-4f6a-a36d-82dadc904495-kube-api-access-lv2br\") pod \"f45f0571-fa49-4f6a-a36d-82dadc904495\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.374503 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-fernet-keys\") pod \"f45f0571-fa49-4f6a-a36d-82dadc904495\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.374598 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-combined-ca-bundle\") pod \"f45f0571-fa49-4f6a-a36d-82dadc904495\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.374641 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-config-data\") pod \"f45f0571-fa49-4f6a-a36d-82dadc904495\" (UID: \"f45f0571-fa49-4f6a-a36d-82dadc904495\") " Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.386716 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45f0571-fa49-4f6a-a36d-82dadc904495-kube-api-access-lv2br" (OuterVolumeSpecName: "kube-api-access-lv2br") pod "f45f0571-fa49-4f6a-a36d-82dadc904495" (UID: "f45f0571-fa49-4f6a-a36d-82dadc904495"). InnerVolumeSpecName "kube-api-access-lv2br". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.389192 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f45f0571-fa49-4f6a-a36d-82dadc904495" (UID: "f45f0571-fa49-4f6a-a36d-82dadc904495"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.419987 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f45f0571-fa49-4f6a-a36d-82dadc904495" (UID: "f45f0571-fa49-4f6a-a36d-82dadc904495"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.444175 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-config-data" (OuterVolumeSpecName: "config-data") pod "f45f0571-fa49-4f6a-a36d-82dadc904495" (UID: "f45f0571-fa49-4f6a-a36d-82dadc904495"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.476691 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv2br\" (UniqueName: \"kubernetes.io/projected/f45f0571-fa49-4f6a-a36d-82dadc904495-kube-api-access-lv2br\") on node \"crc\" DevicePath \"\"" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.476729 4685 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.476742 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.476752 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45f0571-fa49-4f6a-a36d-82dadc904495-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.854960 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413861-26zsj" event={"ID":"f45f0571-fa49-4f6a-a36d-82dadc904495","Type":"ContainerDied","Data":"4ff354e3134557f79822d53991223b0e87cc44c5c9962b63b325c7045025caaf"} Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.855001 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff354e3134557f79822d53991223b0e87cc44c5c9962b63b325c7045025caaf" Dec 04 07:01:05 crc kubenswrapper[4685]: I1204 07:01:05.855062 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413861-26zsj" Dec 04 07:01:10 crc kubenswrapper[4685]: I1204 07:01:10.125517 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:01:10 crc kubenswrapper[4685]: E1204 07:01:10.126316 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:01:24 crc kubenswrapper[4685]: I1204 07:01:24.126705 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:01:24 crc kubenswrapper[4685]: E1204 07:01:24.127367 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:01:38 crc kubenswrapper[4685]: I1204 07:01:38.127451 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:01:38 crc kubenswrapper[4685]: E1204 07:01:38.128311 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:01:53 crc kubenswrapper[4685]: I1204 07:01:53.139595 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:01:53 crc kubenswrapper[4685]: E1204 07:01:53.140336 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:02:07 crc kubenswrapper[4685]: I1204 07:02:07.126964 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:02:07 crc kubenswrapper[4685]: E1204 07:02:07.128360 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:02:19 crc kubenswrapper[4685]: I1204 07:02:19.126702 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:02:19 crc kubenswrapper[4685]: E1204 07:02:19.128920 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:02:32 crc kubenswrapper[4685]: I1204 07:02:32.125977 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:02:32 crc kubenswrapper[4685]: E1204 07:02:32.128346 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:02:47 crc kubenswrapper[4685]: I1204 07:02:47.125930 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:02:47 crc kubenswrapper[4685]: E1204 07:02:47.126687 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:03:02 crc kubenswrapper[4685]: I1204 07:03:02.126223 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:03:02 crc kubenswrapper[4685]: E1204 07:03:02.126990 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:03:16 crc kubenswrapper[4685]: I1204 07:03:16.125999 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:03:16 crc kubenswrapper[4685]: E1204 07:03:16.127163 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:03:30 crc kubenswrapper[4685]: I1204 07:03:30.126071 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:03:30 crc kubenswrapper[4685]: E1204 07:03:30.126802 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:03:41 crc kubenswrapper[4685]: I1204 07:03:41.130806 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:03:41 crc kubenswrapper[4685]: E1204 07:03:41.131862 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:03:54 crc kubenswrapper[4685]: I1204 07:03:54.125808 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:03:54 crc kubenswrapper[4685]: E1204 07:03:54.126859 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:04:06 crc kubenswrapper[4685]: I1204 07:04:06.126189 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:04:06 crc kubenswrapper[4685]: E1204 07:04:06.127079 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:04:17 crc kubenswrapper[4685]: I1204 07:04:17.126658 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:04:17 crc kubenswrapper[4685]: E1204 07:04:17.128275 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:04:32 crc kubenswrapper[4685]: I1204 07:04:32.125329 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:04:32 crc kubenswrapper[4685]: E1204 07:04:32.126155 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:04:44 crc kubenswrapper[4685]: I1204 07:04:44.126495 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:04:45 crc kubenswrapper[4685]: I1204 07:04:45.266501 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"a23eb7765c09a2a6f693470deeb83a7a83b3c8a6bd61e7901d926eabf678683d"} Dec 04 07:07:06 crc kubenswrapper[4685]: I1204 07:07:06.121356 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:07:06 crc kubenswrapper[4685]: I1204 07:07:06.121897 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:07:14 crc kubenswrapper[4685]: I1204 07:07:14.976974 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cl65z"] Dec 04 07:07:14 crc kubenswrapper[4685]: E1204 07:07:14.978259 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45f0571-fa49-4f6a-a36d-82dadc904495" containerName="keystone-cron" Dec 04 07:07:14 crc kubenswrapper[4685]: I1204 07:07:14.978283 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45f0571-fa49-4f6a-a36d-82dadc904495" containerName="keystone-cron" Dec 04 07:07:14 crc kubenswrapper[4685]: I1204 07:07:14.978676 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45f0571-fa49-4f6a-a36d-82dadc904495" containerName="keystone-cron" Dec 04 07:07:14 crc kubenswrapper[4685]: I1204 07:07:14.981213 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:14.987895 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl65z"] Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.110401 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-catalog-content\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.110498 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk5nf\" (UniqueName: \"kubernetes.io/projected/7c3a8d93-96c1-42b5-b43c-de083b3228fa-kube-api-access-dk5nf\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.110615 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-utilities\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.211950 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-utilities\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.212016 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-catalog-content\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.212077 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk5nf\" (UniqueName: \"kubernetes.io/projected/7c3a8d93-96c1-42b5-b43c-de083b3228fa-kube-api-access-dk5nf\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.212486 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-utilities\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.212637 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-catalog-content\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.230642 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk5nf\" (UniqueName: \"kubernetes.io/projected/7c3a8d93-96c1-42b5-b43c-de083b3228fa-kube-api-access-dk5nf\") pod \"redhat-marketplace-cl65z\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.338231 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:15 crc kubenswrapper[4685]: I1204 07:07:15.805934 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl65z"] Dec 04 07:07:16 crc kubenswrapper[4685]: I1204 07:07:16.755773 4685 generic.go:334] "Generic (PLEG): container finished" podID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerID="b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6" exitCode=0 Dec 04 07:07:16 crc kubenswrapper[4685]: I1204 07:07:16.755967 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerDied","Data":"b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6"} Dec 04 07:07:16 crc kubenswrapper[4685]: I1204 07:07:16.756087 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerStarted","Data":"661ae413e5c6bb5667c93185ad2fd566799638901a03140a9c86a0590d8d6bc4"} Dec 04 07:07:16 crc kubenswrapper[4685]: I1204 07:07:16.758600 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 07:07:17 crc kubenswrapper[4685]: I1204 07:07:17.769972 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerStarted","Data":"681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc"} Dec 04 07:07:18 crc kubenswrapper[4685]: I1204 07:07:18.779916 4685 generic.go:334] "Generic (PLEG): container finished" podID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerID="681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc" exitCode=0 Dec 04 07:07:18 crc kubenswrapper[4685]: I1204 07:07:18.779962 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerDied","Data":"681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc"} Dec 04 07:07:19 crc kubenswrapper[4685]: I1204 07:07:19.793823 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerStarted","Data":"4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737"} Dec 04 07:07:19 crc kubenswrapper[4685]: I1204 07:07:19.818529 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cl65z" podStartSLOduration=3.412443369 podStartE2EDuration="5.818508458s" podCreationTimestamp="2025-12-04 07:07:14 +0000 UTC" firstStartedPulling="2025-12-04 07:07:16.758271542 +0000 UTC m=+3373.946502317" lastFinishedPulling="2025-12-04 07:07:19.164336601 +0000 UTC m=+3376.352567406" observedRunningTime="2025-12-04 07:07:19.810968043 +0000 UTC m=+3376.999198838" watchObservedRunningTime="2025-12-04 07:07:19.818508458 +0000 UTC m=+3377.006739243" Dec 04 07:07:25 crc kubenswrapper[4685]: I1204 07:07:25.339286 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:25 crc kubenswrapper[4685]: I1204 07:07:25.340029 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:25 crc kubenswrapper[4685]: I1204 07:07:25.406296 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:25 crc kubenswrapper[4685]: I1204 07:07:25.960287 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:26 crc kubenswrapper[4685]: I1204 07:07:26.034220 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl65z"] Dec 04 07:07:27 crc kubenswrapper[4685]: I1204 07:07:27.903542 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cl65z" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="registry-server" containerID="cri-o://4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737" gracePeriod=2 Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.470228 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.480395 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-catalog-content\") pod \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.480578 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-utilities\") pod \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.480758 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk5nf\" (UniqueName: \"kubernetes.io/projected/7c3a8d93-96c1-42b5-b43c-de083b3228fa-kube-api-access-dk5nf\") pod \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\" (UID: \"7c3a8d93-96c1-42b5-b43c-de083b3228fa\") " Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.481734 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-utilities" (OuterVolumeSpecName: "utilities") pod "7c3a8d93-96c1-42b5-b43c-de083b3228fa" (UID: "7c3a8d93-96c1-42b5-b43c-de083b3228fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.491893 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3a8d93-96c1-42b5-b43c-de083b3228fa-kube-api-access-dk5nf" (OuterVolumeSpecName: "kube-api-access-dk5nf") pod "7c3a8d93-96c1-42b5-b43c-de083b3228fa" (UID: "7c3a8d93-96c1-42b5-b43c-de083b3228fa"). InnerVolumeSpecName "kube-api-access-dk5nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.512637 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c3a8d93-96c1-42b5-b43c-de083b3228fa" (UID: "7c3a8d93-96c1-42b5-b43c-de083b3228fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.583286 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.583327 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk5nf\" (UniqueName: \"kubernetes.io/projected/7c3a8d93-96c1-42b5-b43c-de083b3228fa-kube-api-access-dk5nf\") on node \"crc\" DevicePath \"\"" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.583341 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3a8d93-96c1-42b5-b43c-de083b3228fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.925588 4685 generic.go:334] "Generic (PLEG): container finished" podID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerID="4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737" exitCode=0 Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.925661 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerDied","Data":"4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737"} Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.925686 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cl65z" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.925781 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl65z" event={"ID":"7c3a8d93-96c1-42b5-b43c-de083b3228fa","Type":"ContainerDied","Data":"661ae413e5c6bb5667c93185ad2fd566799638901a03140a9c86a0590d8d6bc4"} Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.925855 4685 scope.go:117] "RemoveContainer" containerID="4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.971690 4685 scope.go:117] "RemoveContainer" containerID="681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc" Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.983331 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl65z"] Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.995573 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl65z"] Dec 04 07:07:28 crc kubenswrapper[4685]: I1204 07:07:28.995630 4685 scope.go:117] "RemoveContainer" containerID="b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.053175 4685 scope.go:117] "RemoveContainer" containerID="4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737" Dec 04 07:07:29 crc kubenswrapper[4685]: E1204 07:07:29.053870 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737\": container with ID starting with 4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737 not found: ID does not exist" containerID="4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.053934 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737"} err="failed to get container status \"4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737\": rpc error: code = NotFound desc = could not find container \"4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737\": container with ID starting with 4fab6849ba4b1eb71ace0a4be0e5508d9c65177553cc368ea836b5c6c8f8c737 not found: ID does not exist" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.053968 4685 scope.go:117] "RemoveContainer" containerID="681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc" Dec 04 07:07:29 crc kubenswrapper[4685]: E1204 07:07:29.054491 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc\": container with ID starting with 681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc not found: ID does not exist" containerID="681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.054533 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc"} err="failed to get container status \"681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc\": rpc error: code = NotFound desc = could not find container \"681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc\": container with ID starting with 681b1af597407bb29fd8fd5d80d9732f9fade49ba13f55d5e992893a5654defc not found: ID does not exist" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.054560 4685 scope.go:117] "RemoveContainer" containerID="b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6" Dec 04 07:07:29 crc kubenswrapper[4685]: E1204 07:07:29.055148 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6\": container with ID starting with b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6 not found: ID does not exist" containerID="b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.055245 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6"} err="failed to get container status \"b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6\": rpc error: code = NotFound desc = could not find container \"b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6\": container with ID starting with b1538c728fd5735244547b65e00f401f9fca7fa4ff2127f40eae7dff8cd798e6 not found: ID does not exist" Dec 04 07:07:29 crc kubenswrapper[4685]: I1204 07:07:29.139038 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" path="/var/lib/kubelet/pods/7c3a8d93-96c1-42b5-b43c-de083b3228fa/volumes" Dec 04 07:07:36 crc kubenswrapper[4685]: I1204 07:07:36.121784 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:07:36 crc kubenswrapper[4685]: I1204 07:07:36.122397 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.121059 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.121663 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.121716 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.122572 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a23eb7765c09a2a6f693470deeb83a7a83b3c8a6bd61e7901d926eabf678683d"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.122643 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://a23eb7765c09a2a6f693470deeb83a7a83b3c8a6bd61e7901d926eabf678683d" gracePeriod=600 Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.323383 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="a23eb7765c09a2a6f693470deeb83a7a83b3c8a6bd61e7901d926eabf678683d" exitCode=0 Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.323442 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"a23eb7765c09a2a6f693470deeb83a7a83b3c8a6bd61e7901d926eabf678683d"} Dec 04 07:08:06 crc kubenswrapper[4685]: I1204 07:08:06.323475 4685 scope.go:117] "RemoveContainer" containerID="1a08a2a6253cd60afbb41222412509b2700ac583f18beb78e9999d3c46175b64" Dec 04 07:08:07 crc kubenswrapper[4685]: I1204 07:08:07.332547 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f"} Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.860903 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s6n77"] Dec 04 07:08:27 crc kubenswrapper[4685]: E1204 07:08:27.861892 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="extract-utilities" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.861907 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="extract-utilities" Dec 04 07:08:27 crc kubenswrapper[4685]: E1204 07:08:27.861939 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="extract-content" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.861949 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="extract-content" Dec 04 07:08:27 crc kubenswrapper[4685]: E1204 07:08:27.861976 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="registry-server" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.861985 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="registry-server" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.862218 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3a8d93-96c1-42b5-b43c-de083b3228fa" containerName="registry-server" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.863804 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.899138 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s6n77"] Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.928495 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-catalog-content\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.928867 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-utilities\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:27 crc kubenswrapper[4685]: I1204 07:08:27.928961 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlz5m\" (UniqueName: \"kubernetes.io/projected/9f3bbf03-53e3-4153-94f2-8fd6299936a2-kube-api-access-rlz5m\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.031420 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-catalog-content\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.031828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-utilities\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.031866 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlz5m\" (UniqueName: \"kubernetes.io/projected/9f3bbf03-53e3-4153-94f2-8fd6299936a2-kube-api-access-rlz5m\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.032060 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-catalog-content\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.032447 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-utilities\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.057257 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlz5m\" (UniqueName: \"kubernetes.io/projected/9f3bbf03-53e3-4153-94f2-8fd6299936a2-kube-api-access-rlz5m\") pod \"certified-operators-s6n77\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.191809 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:28 crc kubenswrapper[4685]: I1204 07:08:28.669349 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s6n77"] Dec 04 07:08:29 crc kubenswrapper[4685]: I1204 07:08:29.570364 4685 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerID="d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036" exitCode=0 Dec 04 07:08:29 crc kubenswrapper[4685]: I1204 07:08:29.570483 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerDied","Data":"d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036"} Dec 04 07:08:29 crc kubenswrapper[4685]: I1204 07:08:29.570733 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerStarted","Data":"c4fa26242b4dd86c96db6e90ed102ff844ae2cfe933cb2036244bf128e0ae00a"} Dec 04 07:08:30 crc kubenswrapper[4685]: I1204 07:08:30.582723 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerStarted","Data":"5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5"} Dec 04 07:08:31 crc kubenswrapper[4685]: I1204 07:08:31.596440 4685 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerID="5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5" exitCode=0 Dec 04 07:08:31 crc kubenswrapper[4685]: I1204 07:08:31.596504 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerDied","Data":"5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5"} Dec 04 07:08:32 crc kubenswrapper[4685]: I1204 07:08:32.610427 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerStarted","Data":"2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2"} Dec 04 07:08:32 crc kubenswrapper[4685]: I1204 07:08:32.645732 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s6n77" podStartSLOduration=2.942131118 podStartE2EDuration="5.645711429s" podCreationTimestamp="2025-12-04 07:08:27 +0000 UTC" firstStartedPulling="2025-12-04 07:08:29.572934473 +0000 UTC m=+3446.761165258" lastFinishedPulling="2025-12-04 07:08:32.276514784 +0000 UTC m=+3449.464745569" observedRunningTime="2025-12-04 07:08:32.62871447 +0000 UTC m=+3449.816945265" watchObservedRunningTime="2025-12-04 07:08:32.645711429 +0000 UTC m=+3449.833942204" Dec 04 07:08:38 crc kubenswrapper[4685]: I1204 07:08:38.192869 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:38 crc kubenswrapper[4685]: I1204 07:08:38.193668 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:38 crc kubenswrapper[4685]: I1204 07:08:38.263574 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:38 crc kubenswrapper[4685]: I1204 07:08:38.742834 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:38 crc kubenswrapper[4685]: I1204 07:08:38.799573 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s6n77"] Dec 04 07:08:40 crc kubenswrapper[4685]: I1204 07:08:40.690564 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s6n77" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="registry-server" containerID="cri-o://2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2" gracePeriod=2 Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.155625 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.223259 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlz5m\" (UniqueName: \"kubernetes.io/projected/9f3bbf03-53e3-4153-94f2-8fd6299936a2-kube-api-access-rlz5m\") pod \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.223380 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-utilities\") pod \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.223462 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-catalog-content\") pod \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\" (UID: \"9f3bbf03-53e3-4153-94f2-8fd6299936a2\") " Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.225309 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-utilities" (OuterVolumeSpecName: "utilities") pod "9f3bbf03-53e3-4153-94f2-8fd6299936a2" (UID: "9f3bbf03-53e3-4153-94f2-8fd6299936a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.231931 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f3bbf03-53e3-4153-94f2-8fd6299936a2-kube-api-access-rlz5m" (OuterVolumeSpecName: "kube-api-access-rlz5m") pod "9f3bbf03-53e3-4153-94f2-8fd6299936a2" (UID: "9f3bbf03-53e3-4153-94f2-8fd6299936a2"). InnerVolumeSpecName "kube-api-access-rlz5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.277951 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f3bbf03-53e3-4153-94f2-8fd6299936a2" (UID: "9f3bbf03-53e3-4153-94f2-8fd6299936a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.327991 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlz5m\" (UniqueName: \"kubernetes.io/projected/9f3bbf03-53e3-4153-94f2-8fd6299936a2-kube-api-access-rlz5m\") on node \"crc\" DevicePath \"\"" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.328037 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.328057 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbf03-53e3-4153-94f2-8fd6299936a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.721290 4685 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerID="2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2" exitCode=0 Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.722273 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerDied","Data":"2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2"} Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.722399 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6n77" event={"ID":"9f3bbf03-53e3-4153-94f2-8fd6299936a2","Type":"ContainerDied","Data":"c4fa26242b4dd86c96db6e90ed102ff844ae2cfe933cb2036244bf128e0ae00a"} Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.722515 4685 scope.go:117] "RemoveContainer" containerID="2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.722806 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6n77" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.761700 4685 scope.go:117] "RemoveContainer" containerID="5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.780950 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s6n77"] Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.787052 4685 scope.go:117] "RemoveContainer" containerID="d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.790061 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s6n77"] Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.841708 4685 scope.go:117] "RemoveContainer" containerID="2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2" Dec 04 07:08:41 crc kubenswrapper[4685]: E1204 07:08:41.842174 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2\": container with ID starting with 2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2 not found: ID does not exist" containerID="2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.842214 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2"} err="failed to get container status \"2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2\": rpc error: code = NotFound desc = could not find container \"2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2\": container with ID starting with 2df4986b6b720993cd51e90b26d18b556b2126c2b6a946aba13ecf97859f35e2 not found: ID does not exist" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.842238 4685 scope.go:117] "RemoveContainer" containerID="5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5" Dec 04 07:08:41 crc kubenswrapper[4685]: E1204 07:08:41.842537 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5\": container with ID starting with 5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5 not found: ID does not exist" containerID="5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.842584 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5"} err="failed to get container status \"5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5\": rpc error: code = NotFound desc = could not find container \"5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5\": container with ID starting with 5f9b1b5ec023e4a4fe20d5e4a72d83c6b6305e81eca7c1893f1c954dc4b161f5 not found: ID does not exist" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.842617 4685 scope.go:117] "RemoveContainer" containerID="d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036" Dec 04 07:08:41 crc kubenswrapper[4685]: E1204 07:08:41.842856 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036\": container with ID starting with d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036 not found: ID does not exist" containerID="d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036" Dec 04 07:08:41 crc kubenswrapper[4685]: I1204 07:08:41.842881 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036"} err="failed to get container status \"d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036\": rpc error: code = NotFound desc = could not find container \"d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036\": container with ID starting with d6eec3e7ccf246ffa0bfb923c8847f5a9ae8b7ea5cfe82110b4186dbe935d036 not found: ID does not exist" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.138913 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" path="/var/lib/kubelet/pods/9f3bbf03-53e3-4153-94f2-8fd6299936a2/volumes" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.927592 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t24ng"] Dec 04 07:08:43 crc kubenswrapper[4685]: E1204 07:08:43.928049 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="registry-server" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.928071 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="registry-server" Dec 04 07:08:43 crc kubenswrapper[4685]: E1204 07:08:43.928086 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="extract-content" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.928095 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="extract-content" Dec 04 07:08:43 crc kubenswrapper[4685]: E1204 07:08:43.928147 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="extract-utilities" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.928156 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="extract-utilities" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.928438 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3bbf03-53e3-4153-94f2-8fd6299936a2" containerName="registry-server" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.930080 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.943810 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t24ng"] Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.978902 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-utilities\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.978969 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-catalog-content\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:43 crc kubenswrapper[4685]: I1204 07:08:43.979378 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96wzx\" (UniqueName: \"kubernetes.io/projected/01aeee99-a63e-45ba-bc6e-8564612baa47-kube-api-access-96wzx\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.082081 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-utilities\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.082369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-catalog-content\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.082564 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96wzx\" (UniqueName: \"kubernetes.io/projected/01aeee99-a63e-45ba-bc6e-8564612baa47-kube-api-access-96wzx\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.083057 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-catalog-content\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.083261 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-utilities\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.111061 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96wzx\" (UniqueName: \"kubernetes.io/projected/01aeee99-a63e-45ba-bc6e-8564612baa47-kube-api-access-96wzx\") pod \"community-operators-t24ng\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.266614 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:44 crc kubenswrapper[4685]: I1204 07:08:44.802172 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t24ng"] Dec 04 07:08:45 crc kubenswrapper[4685]: I1204 07:08:45.756268 4685 generic.go:334] "Generic (PLEG): container finished" podID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerID="bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb" exitCode=0 Dec 04 07:08:45 crc kubenswrapper[4685]: I1204 07:08:45.756348 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerDied","Data":"bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb"} Dec 04 07:08:45 crc kubenswrapper[4685]: I1204 07:08:45.756621 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerStarted","Data":"7a047e07c5f43a5aa2e48b1e98d8d99b4e349de56284fdebc6825d84280ed3ec"} Dec 04 07:08:46 crc kubenswrapper[4685]: I1204 07:08:46.769391 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerStarted","Data":"f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53"} Dec 04 07:08:47 crc kubenswrapper[4685]: I1204 07:08:47.783366 4685 generic.go:334] "Generic (PLEG): container finished" podID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerID="f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53" exitCode=0 Dec 04 07:08:47 crc kubenswrapper[4685]: I1204 07:08:47.783421 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerDied","Data":"f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53"} Dec 04 07:08:48 crc kubenswrapper[4685]: I1204 07:08:48.803854 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerStarted","Data":"4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1"} Dec 04 07:08:48 crc kubenswrapper[4685]: I1204 07:08:48.824216 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t24ng" podStartSLOduration=3.365100434 podStartE2EDuration="5.824195406s" podCreationTimestamp="2025-12-04 07:08:43 +0000 UTC" firstStartedPulling="2025-12-04 07:08:45.759003606 +0000 UTC m=+3462.947234381" lastFinishedPulling="2025-12-04 07:08:48.218098578 +0000 UTC m=+3465.406329353" observedRunningTime="2025-12-04 07:08:48.819075437 +0000 UTC m=+3466.007306212" watchObservedRunningTime="2025-12-04 07:08:48.824195406 +0000 UTC m=+3466.012426171" Dec 04 07:08:54 crc kubenswrapper[4685]: I1204 07:08:54.267162 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:54 crc kubenswrapper[4685]: I1204 07:08:54.267827 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:54 crc kubenswrapper[4685]: I1204 07:08:54.328648 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:54 crc kubenswrapper[4685]: I1204 07:08:54.916475 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:54 crc kubenswrapper[4685]: I1204 07:08:54.962062 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t24ng"] Dec 04 07:08:56 crc kubenswrapper[4685]: I1204 07:08:56.891389 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t24ng" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="registry-server" containerID="cri-o://4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1" gracePeriod=2 Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.883710 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.903130 4685 generic.go:334] "Generic (PLEG): container finished" podID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerID="4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1" exitCode=0 Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.903172 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t24ng" Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.903193 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerDied","Data":"4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1"} Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.903365 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t24ng" event={"ID":"01aeee99-a63e-45ba-bc6e-8564612baa47","Type":"ContainerDied","Data":"7a047e07c5f43a5aa2e48b1e98d8d99b4e349de56284fdebc6825d84280ed3ec"} Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.903503 4685 scope.go:117] "RemoveContainer" containerID="4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1" Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.934381 4685 scope.go:117] "RemoveContainer" containerID="f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53" Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.958242 4685 scope.go:117] "RemoveContainer" containerID="bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb" Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.970763 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-utilities\") pod \"01aeee99-a63e-45ba-bc6e-8564612baa47\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.970899 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-catalog-content\") pod \"01aeee99-a63e-45ba-bc6e-8564612baa47\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.970987 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96wzx\" (UniqueName: \"kubernetes.io/projected/01aeee99-a63e-45ba-bc6e-8564612baa47-kube-api-access-96wzx\") pod \"01aeee99-a63e-45ba-bc6e-8564612baa47\" (UID: \"01aeee99-a63e-45ba-bc6e-8564612baa47\") " Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.971801 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-utilities" (OuterVolumeSpecName: "utilities") pod "01aeee99-a63e-45ba-bc6e-8564612baa47" (UID: "01aeee99-a63e-45ba-bc6e-8564612baa47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:08:57 crc kubenswrapper[4685]: I1204 07:08:57.981699 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01aeee99-a63e-45ba-bc6e-8564612baa47-kube-api-access-96wzx" (OuterVolumeSpecName: "kube-api-access-96wzx") pod "01aeee99-a63e-45ba-bc6e-8564612baa47" (UID: "01aeee99-a63e-45ba-bc6e-8564612baa47"). InnerVolumeSpecName "kube-api-access-96wzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.024785 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01aeee99-a63e-45ba-bc6e-8564612baa47" (UID: "01aeee99-a63e-45ba-bc6e-8564612baa47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.056081 4685 scope.go:117] "RemoveContainer" containerID="4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1" Dec 04 07:08:58 crc kubenswrapper[4685]: E1204 07:08:58.056575 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1\": container with ID starting with 4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1 not found: ID does not exist" containerID="4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.056622 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1"} err="failed to get container status \"4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1\": rpc error: code = NotFound desc = could not find container \"4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1\": container with ID starting with 4029818e28662f1cdb08dbb3f428a4f1ac439e9fcc1061c23934e5c9db0012e1 not found: ID does not exist" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.056648 4685 scope.go:117] "RemoveContainer" containerID="f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53" Dec 04 07:08:58 crc kubenswrapper[4685]: E1204 07:08:58.056975 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53\": container with ID starting with f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53 not found: ID does not exist" containerID="f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.057016 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53"} err="failed to get container status \"f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53\": rpc error: code = NotFound desc = could not find container \"f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53\": container with ID starting with f55d807158461ed152d974ac7f8c77f80903035fa528c3dd081f3d65682e2a53 not found: ID does not exist" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.057041 4685 scope.go:117] "RemoveContainer" containerID="bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb" Dec 04 07:08:58 crc kubenswrapper[4685]: E1204 07:08:58.057721 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb\": container with ID starting with bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb not found: ID does not exist" containerID="bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.057754 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb"} err="failed to get container status \"bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb\": rpc error: code = NotFound desc = could not find container \"bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb\": container with ID starting with bd241e257f4f61ea966880861b8a9adb931e7366991106376b07f6a2718170bb not found: ID does not exist" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.073469 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.073502 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01aeee99-a63e-45ba-bc6e-8564612baa47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.073518 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96wzx\" (UniqueName: \"kubernetes.io/projected/01aeee99-a63e-45ba-bc6e-8564612baa47-kube-api-access-96wzx\") on node \"crc\" DevicePath \"\"" Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.248524 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t24ng"] Dec 04 07:08:58 crc kubenswrapper[4685]: I1204 07:08:58.258110 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t24ng"] Dec 04 07:08:59 crc kubenswrapper[4685]: I1204 07:08:59.140597 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" path="/var/lib/kubelet/pods/01aeee99-a63e-45ba-bc6e-8564612baa47/volumes" Dec 04 07:10:06 crc kubenswrapper[4685]: I1204 07:10:06.121365 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:10:06 crc kubenswrapper[4685]: I1204 07:10:06.122026 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:10:36 crc kubenswrapper[4685]: I1204 07:10:36.120944 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:10:36 crc kubenswrapper[4685]: I1204 07:10:36.121636 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:11:06 crc kubenswrapper[4685]: I1204 07:11:06.121335 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:11:06 crc kubenswrapper[4685]: I1204 07:11:06.122943 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:11:06 crc kubenswrapper[4685]: I1204 07:11:06.123125 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 07:11:06 crc kubenswrapper[4685]: I1204 07:11:06.124185 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 07:11:06 crc kubenswrapper[4685]: I1204 07:11:06.124380 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" gracePeriod=600 Dec 04 07:11:06 crc kubenswrapper[4685]: E1204 07:11:06.254150 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:11:07 crc kubenswrapper[4685]: I1204 07:11:07.187524 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" exitCode=0 Dec 04 07:11:07 crc kubenswrapper[4685]: I1204 07:11:07.187580 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f"} Dec 04 07:11:07 crc kubenswrapper[4685]: I1204 07:11:07.187840 4685 scope.go:117] "RemoveContainer" containerID="a23eb7765c09a2a6f693470deeb83a7a83b3c8a6bd61e7901d926eabf678683d" Dec 04 07:11:07 crc kubenswrapper[4685]: I1204 07:11:07.188578 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:11:07 crc kubenswrapper[4685]: E1204 07:11:07.189016 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:11:13 crc kubenswrapper[4685]: I1204 07:11:13.250063 4685 generic.go:334] "Generic (PLEG): container finished" podID="c8fff46b-61e1-48da-a047-9ace06e11177" containerID="f9039625c021671b160c25e306848ed8932c307e304aef00bcb5172eb7ddff90" exitCode=0 Dec 04 07:11:13 crc kubenswrapper[4685]: I1204 07:11:13.250144 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c8fff46b-61e1-48da-a047-9ace06e11177","Type":"ContainerDied","Data":"f9039625c021671b160c25e306848ed8932c307e304aef00bcb5172eb7ddff90"} Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.686903 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736206 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd4b4\" (UniqueName: \"kubernetes.io/projected/c8fff46b-61e1-48da-a047-9ace06e11177-kube-api-access-pd4b4\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736256 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ssh-key\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736275 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736395 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-config-data\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736446 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-workdir\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736476 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-temporary\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736510 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config-secret\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736545 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ca-certs\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.736564 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config\") pod \"c8fff46b-61e1-48da-a047-9ace06e11177\" (UID: \"c8fff46b-61e1-48da-a047-9ace06e11177\") " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.737805 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.739153 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-config-data" (OuterVolumeSpecName: "config-data") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.744368 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.748759 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.748786 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8fff46b-61e1-48da-a047-9ace06e11177-kube-api-access-pd4b4" (OuterVolumeSpecName: "kube-api-access-pd4b4") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "kube-api-access-pd4b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.766256 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.769501 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.782621 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.793447 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c8fff46b-61e1-48da-a047-9ace06e11177" (UID: "c8fff46b-61e1-48da-a047-9ace06e11177"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839021 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839056 4685 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839070 4685 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c8fff46b-61e1-48da-a047-9ace06e11177-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839081 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839090 4685 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839097 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8fff46b-61e1-48da-a047-9ace06e11177-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839105 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd4b4\" (UniqueName: \"kubernetes.io/projected/c8fff46b-61e1-48da-a047-9ace06e11177-kube-api-access-pd4b4\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839113 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8fff46b-61e1-48da-a047-9ace06e11177-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.839150 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.858045 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 04 07:11:14 crc kubenswrapper[4685]: I1204 07:11:14.940610 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:15 crc kubenswrapper[4685]: I1204 07:11:15.270647 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c8fff46b-61e1-48da-a047-9ace06e11177","Type":"ContainerDied","Data":"f639b34c380e5e7348937a7306792d23c7f8b8f56c678f6a5cb013e3affdeabc"} Dec 04 07:11:15 crc kubenswrapper[4685]: I1204 07:11:15.270683 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f639b34c380e5e7348937a7306792d23c7f8b8f56c678f6a5cb013e3affdeabc" Dec 04 07:11:15 crc kubenswrapper[4685]: I1204 07:11:15.270714 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 07:11:22 crc kubenswrapper[4685]: I1204 07:11:22.126355 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:11:22 crc kubenswrapper[4685]: E1204 07:11:22.127728 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.658753 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 07:11:26 crc kubenswrapper[4685]: E1204 07:11:26.660222 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="extract-utilities" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.660255 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="extract-utilities" Dec 04 07:11:26 crc kubenswrapper[4685]: E1204 07:11:26.660301 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fff46b-61e1-48da-a047-9ace06e11177" containerName="tempest-tests-tempest-tests-runner" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.660319 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fff46b-61e1-48da-a047-9ace06e11177" containerName="tempest-tests-tempest-tests-runner" Dec 04 07:11:26 crc kubenswrapper[4685]: E1204 07:11:26.660368 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="extract-content" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.660386 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="extract-content" Dec 04 07:11:26 crc kubenswrapper[4685]: E1204 07:11:26.660464 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="registry-server" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.660484 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="registry-server" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.660904 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="01aeee99-a63e-45ba-bc6e-8564612baa47" containerName="registry-server" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.660967 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8fff46b-61e1-48da-a047-9ace06e11177" containerName="tempest-tests-tempest-tests-runner" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.662278 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.664357 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fdj2s" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.667453 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.779665 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brjcd\" (UniqueName: \"kubernetes.io/projected/98d35031-8372-42b2-9b45-e068ff8b4c08-kube-api-access-brjcd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.779731 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.881857 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brjcd\" (UniqueName: \"kubernetes.io/projected/98d35031-8372-42b2-9b45-e068ff8b4c08-kube-api-access-brjcd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.882449 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.884168 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.923310 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.926819 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brjcd\" (UniqueName: \"kubernetes.io/projected/98d35031-8372-42b2-9b45-e068ff8b4c08-kube-api-access-brjcd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"98d35031-8372-42b2-9b45-e068ff8b4c08\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:26 crc kubenswrapper[4685]: I1204 07:11:26.990914 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 07:11:27 crc kubenswrapper[4685]: I1204 07:11:27.480047 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 07:11:28 crc kubenswrapper[4685]: I1204 07:11:28.402072 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"98d35031-8372-42b2-9b45-e068ff8b4c08","Type":"ContainerStarted","Data":"64eef50f076d5dfd056dcd36f815627261b429cd0a39c7f87f3e5d693b98f52f"} Dec 04 07:11:29 crc kubenswrapper[4685]: I1204 07:11:29.417830 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"98d35031-8372-42b2-9b45-e068ff8b4c08","Type":"ContainerStarted","Data":"509f231ca1869f275ec7aa5029050ba402ecf1d75d40f0399322dc1ab6116d37"} Dec 04 07:11:37 crc kubenswrapper[4685]: I1204 07:11:37.126015 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:11:37 crc kubenswrapper[4685]: E1204 07:11:37.126879 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.663125 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=11.773162487 podStartE2EDuration="12.663105498s" podCreationTimestamp="2025-12-04 07:11:26 +0000 UTC" firstStartedPulling="2025-12-04 07:11:27.478769644 +0000 UTC m=+3624.667000449" lastFinishedPulling="2025-12-04 07:11:28.368712675 +0000 UTC m=+3625.556943460" observedRunningTime="2025-12-04 07:11:29.437756463 +0000 UTC m=+3626.625987238" watchObservedRunningTime="2025-12-04 07:11:38.663105498 +0000 UTC m=+3635.851336273" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.675556 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2tdhq"] Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.679870 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.694522 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2tdhq"] Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.832486 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-utilities\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.832805 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5qb9\" (UniqueName: \"kubernetes.io/projected/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-kube-api-access-h5qb9\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.832921 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-catalog-content\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.934436 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5qb9\" (UniqueName: \"kubernetes.io/projected/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-kube-api-access-h5qb9\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.934510 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-catalog-content\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.934608 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-utilities\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.935204 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-utilities\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.935258 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-catalog-content\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:38 crc kubenswrapper[4685]: I1204 07:11:38.959250 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5qb9\" (UniqueName: \"kubernetes.io/projected/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-kube-api-access-h5qb9\") pod \"redhat-operators-2tdhq\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:39 crc kubenswrapper[4685]: I1204 07:11:39.008994 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:39 crc kubenswrapper[4685]: I1204 07:11:39.303867 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2tdhq"] Dec 04 07:11:39 crc kubenswrapper[4685]: I1204 07:11:39.538790 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerID="8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833" exitCode=0 Dec 04 07:11:39 crc kubenswrapper[4685]: I1204 07:11:39.538834 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerDied","Data":"8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833"} Dec 04 07:11:39 crc kubenswrapper[4685]: I1204 07:11:39.538866 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerStarted","Data":"145e448aedd6021414a63ea335c83be3eba72a0fe9779b4ae15d87aedfc3ddbb"} Dec 04 07:11:40 crc kubenswrapper[4685]: I1204 07:11:40.551671 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerStarted","Data":"b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252"} Dec 04 07:11:41 crc kubenswrapper[4685]: I1204 07:11:41.572966 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerID="b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252" exitCode=0 Dec 04 07:11:41 crc kubenswrapper[4685]: I1204 07:11:41.573051 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerDied","Data":"b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252"} Dec 04 07:11:42 crc kubenswrapper[4685]: I1204 07:11:42.585668 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerStarted","Data":"de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6"} Dec 04 07:11:42 crc kubenswrapper[4685]: I1204 07:11:42.613636 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2tdhq" podStartSLOduration=2.130610201 podStartE2EDuration="4.613616406s" podCreationTimestamp="2025-12-04 07:11:38 +0000 UTC" firstStartedPulling="2025-12-04 07:11:39.540740007 +0000 UTC m=+3636.728970782" lastFinishedPulling="2025-12-04 07:11:42.023746172 +0000 UTC m=+3639.211976987" observedRunningTime="2025-12-04 07:11:42.610905501 +0000 UTC m=+3639.799136286" watchObservedRunningTime="2025-12-04 07:11:42.613616406 +0000 UTC m=+3639.801847191" Dec 04 07:11:49 crc kubenswrapper[4685]: I1204 07:11:49.009511 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:49 crc kubenswrapper[4685]: I1204 07:11:49.010004 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:49 crc kubenswrapper[4685]: I1204 07:11:49.057776 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:49 crc kubenswrapper[4685]: I1204 07:11:49.691179 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:50 crc kubenswrapper[4685]: I1204 07:11:50.657705 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2tdhq"] Dec 04 07:11:51 crc kubenswrapper[4685]: I1204 07:11:51.670541 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2tdhq" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="registry-server" containerID="cri-o://de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6" gracePeriod=2 Dec 04 07:11:52 crc kubenswrapper[4685]: I1204 07:11:52.126324 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:11:52 crc kubenswrapper[4685]: E1204 07:11:52.127332 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.271814 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.358007 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-catalog-content\") pod \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.358087 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-utilities\") pod \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.358166 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5qb9\" (UniqueName: \"kubernetes.io/projected/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-kube-api-access-h5qb9\") pod \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\" (UID: \"ac116d3c-5b0f-4f0f-a9a9-078bf1320646\") " Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.360202 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-utilities" (OuterVolumeSpecName: "utilities") pod "ac116d3c-5b0f-4f0f-a9a9-078bf1320646" (UID: "ac116d3c-5b0f-4f0f-a9a9-078bf1320646"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.368076 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-kube-api-access-h5qb9" (OuterVolumeSpecName: "kube-api-access-h5qb9") pod "ac116d3c-5b0f-4f0f-a9a9-078bf1320646" (UID: "ac116d3c-5b0f-4f0f-a9a9-078bf1320646"). InnerVolumeSpecName "kube-api-access-h5qb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.460111 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.460151 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5qb9\" (UniqueName: \"kubernetes.io/projected/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-kube-api-access-h5qb9\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.468418 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac116d3c-5b0f-4f0f-a9a9-078bf1320646" (UID: "ac116d3c-5b0f-4f0f-a9a9-078bf1320646"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.562707 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac116d3c-5b0f-4f0f-a9a9-078bf1320646-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.686161 4685 generic.go:334] "Generic (PLEG): container finished" podID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerID="de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6" exitCode=0 Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.686199 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerDied","Data":"de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6"} Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.686236 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2tdhq" event={"ID":"ac116d3c-5b0f-4f0f-a9a9-078bf1320646","Type":"ContainerDied","Data":"145e448aedd6021414a63ea335c83be3eba72a0fe9779b4ae15d87aedfc3ddbb"} Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.686253 4685 scope.go:117] "RemoveContainer" containerID="de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.686363 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2tdhq" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.713259 4685 scope.go:117] "RemoveContainer" containerID="b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.721868 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2tdhq"] Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.734027 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2tdhq"] Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.748289 4685 scope.go:117] "RemoveContainer" containerID="8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.778504 4685 scope.go:117] "RemoveContainer" containerID="de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6" Dec 04 07:11:53 crc kubenswrapper[4685]: E1204 07:11:53.778901 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6\": container with ID starting with de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6 not found: ID does not exist" containerID="de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.778935 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6"} err="failed to get container status \"de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6\": rpc error: code = NotFound desc = could not find container \"de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6\": container with ID starting with de7cc83d3e5086db3ef9627d5c577b1748c43ccc063045b018825f9cc0f479d6 not found: ID does not exist" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.778955 4685 scope.go:117] "RemoveContainer" containerID="b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252" Dec 04 07:11:53 crc kubenswrapper[4685]: E1204 07:11:53.779268 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252\": container with ID starting with b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252 not found: ID does not exist" containerID="b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.779290 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252"} err="failed to get container status \"b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252\": rpc error: code = NotFound desc = could not find container \"b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252\": container with ID starting with b93fd9496c0f2bca229921de1a3308f1b3876e5880ff69b00f0694fd8dab4252 not found: ID does not exist" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.779303 4685 scope.go:117] "RemoveContainer" containerID="8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833" Dec 04 07:11:53 crc kubenswrapper[4685]: E1204 07:11:53.779753 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833\": container with ID starting with 8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833 not found: ID does not exist" containerID="8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833" Dec 04 07:11:53 crc kubenswrapper[4685]: I1204 07:11:53.779786 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833"} err="failed to get container status \"8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833\": rpc error: code = NotFound desc = could not find container \"8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833\": container with ID starting with 8744a7089d9cbde0f374473dd2404aa7124f0bc35e30586a3624bbff7763a833 not found: ID does not exist" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.135829 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" path="/var/lib/kubelet/pods/ac116d3c-5b0f-4f0f-a9a9-078bf1320646/volumes" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.155887 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v225f/must-gather-rtvxw"] Dec 04 07:11:55 crc kubenswrapper[4685]: E1204 07:11:55.156370 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="extract-content" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.156393 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="extract-content" Dec 04 07:11:55 crc kubenswrapper[4685]: E1204 07:11:55.156436 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="extract-utilities" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.156447 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="extract-utilities" Dec 04 07:11:55 crc kubenswrapper[4685]: E1204 07:11:55.156485 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="registry-server" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.156495 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="registry-server" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.156732 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac116d3c-5b0f-4f0f-a9a9-078bf1320646" containerName="registry-server" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.157982 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.159469 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v225f"/"kube-root-ca.crt" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.159669 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-v225f"/"default-dockercfg-7sppz" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.160038 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v225f"/"openshift-service-ca.crt" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.167142 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v225f/must-gather-rtvxw"] Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.303933 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tglfm\" (UniqueName: \"kubernetes.io/projected/7b59e18d-45b2-43e3-bdac-f28ac18243c6-kube-api-access-tglfm\") pod \"must-gather-rtvxw\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.303996 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7b59e18d-45b2-43e3-bdac-f28ac18243c6-must-gather-output\") pod \"must-gather-rtvxw\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.405532 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tglfm\" (UniqueName: \"kubernetes.io/projected/7b59e18d-45b2-43e3-bdac-f28ac18243c6-kube-api-access-tglfm\") pod \"must-gather-rtvxw\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.405589 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7b59e18d-45b2-43e3-bdac-f28ac18243c6-must-gather-output\") pod \"must-gather-rtvxw\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.406158 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7b59e18d-45b2-43e3-bdac-f28ac18243c6-must-gather-output\") pod \"must-gather-rtvxw\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.424192 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tglfm\" (UniqueName: \"kubernetes.io/projected/7b59e18d-45b2-43e3-bdac-f28ac18243c6-kube-api-access-tglfm\") pod \"must-gather-rtvxw\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.478768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:11:55 crc kubenswrapper[4685]: I1204 07:11:55.987770 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v225f/must-gather-rtvxw"] Dec 04 07:11:56 crc kubenswrapper[4685]: I1204 07:11:56.717043 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/must-gather-rtvxw" event={"ID":"7b59e18d-45b2-43e3-bdac-f28ac18243c6","Type":"ContainerStarted","Data":"d37659c0e9d7ed9d2a4b1620b77fbda1cc1f3b6fb9cb0a1a63496364c934740b"} Dec 04 07:12:00 crc kubenswrapper[4685]: I1204 07:12:00.763498 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/must-gather-rtvxw" event={"ID":"7b59e18d-45b2-43e3-bdac-f28ac18243c6","Type":"ContainerStarted","Data":"53f81f0bc8f98fd95e9961b588f4e6e785e7c13410d3ff5c1a20cb309d72a38b"} Dec 04 07:12:00 crc kubenswrapper[4685]: I1204 07:12:00.763977 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/must-gather-rtvxw" event={"ID":"7b59e18d-45b2-43e3-bdac-f28ac18243c6","Type":"ContainerStarted","Data":"6e305812b42877dd9791806a0cb6987062bf480adc82261425a4459316de1574"} Dec 04 07:12:00 crc kubenswrapper[4685]: I1204 07:12:00.785460 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v225f/must-gather-rtvxw" podStartSLOduration=2.055130465 podStartE2EDuration="5.785441869s" podCreationTimestamp="2025-12-04 07:11:55 +0000 UTC" firstStartedPulling="2025-12-04 07:11:55.986481661 +0000 UTC m=+3653.174712476" lastFinishedPulling="2025-12-04 07:11:59.716793105 +0000 UTC m=+3656.905023880" observedRunningTime="2025-12-04 07:12:00.77998961 +0000 UTC m=+3657.968220395" watchObservedRunningTime="2025-12-04 07:12:00.785441869 +0000 UTC m=+3657.973672654" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.127331 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:12:03 crc kubenswrapper[4685]: E1204 07:12:03.128097 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.583033 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v225f/crc-debug-7bgqz"] Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.585653 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.676358 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/faf74226-3731-4b3a-9479-2289c54ac84d-host\") pod \"crc-debug-7bgqz\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.676615 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b54t\" (UniqueName: \"kubernetes.io/projected/faf74226-3731-4b3a-9479-2289c54ac84d-kube-api-access-5b54t\") pod \"crc-debug-7bgqz\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.778865 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/faf74226-3731-4b3a-9479-2289c54ac84d-host\") pod \"crc-debug-7bgqz\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.778961 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b54t\" (UniqueName: \"kubernetes.io/projected/faf74226-3731-4b3a-9479-2289c54ac84d-kube-api-access-5b54t\") pod \"crc-debug-7bgqz\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.779012 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/faf74226-3731-4b3a-9479-2289c54ac84d-host\") pod \"crc-debug-7bgqz\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.803168 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b54t\" (UniqueName: \"kubernetes.io/projected/faf74226-3731-4b3a-9479-2289c54ac84d-kube-api-access-5b54t\") pod \"crc-debug-7bgqz\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: I1204 07:12:03.907322 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:03 crc kubenswrapper[4685]: W1204 07:12:03.944046 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaf74226_3731_4b3a_9479_2289c54ac84d.slice/crio-458539a6cfefed32fc10e589e11ed1dc7aff0f07914f57be42ec8d2001342f1d WatchSource:0}: Error finding container 458539a6cfefed32fc10e589e11ed1dc7aff0f07914f57be42ec8d2001342f1d: Status 404 returned error can't find the container with id 458539a6cfefed32fc10e589e11ed1dc7aff0f07914f57be42ec8d2001342f1d Dec 04 07:12:04 crc kubenswrapper[4685]: I1204 07:12:04.820154 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-7bgqz" event={"ID":"faf74226-3731-4b3a-9479-2289c54ac84d","Type":"ContainerStarted","Data":"458539a6cfefed32fc10e589e11ed1dc7aff0f07914f57be42ec8d2001342f1d"} Dec 04 07:12:14 crc kubenswrapper[4685]: I1204 07:12:14.912981 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-7bgqz" event={"ID":"faf74226-3731-4b3a-9479-2289c54ac84d","Type":"ContainerStarted","Data":"535c83d81f3e45b7c9306f497e71da4e81f887b25194eb5f092063495003af4a"} Dec 04 07:12:14 crc kubenswrapper[4685]: I1204 07:12:14.934627 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v225f/crc-debug-7bgqz" podStartSLOduration=1.771888704 podStartE2EDuration="11.934605315s" podCreationTimestamp="2025-12-04 07:12:03 +0000 UTC" firstStartedPulling="2025-12-04 07:12:03.954383904 +0000 UTC m=+3661.142614679" lastFinishedPulling="2025-12-04 07:12:14.117100515 +0000 UTC m=+3671.305331290" observedRunningTime="2025-12-04 07:12:14.92447328 +0000 UTC m=+3672.112704065" watchObservedRunningTime="2025-12-04 07:12:14.934605315 +0000 UTC m=+3672.122836090" Dec 04 07:12:17 crc kubenswrapper[4685]: I1204 07:12:17.125425 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:12:17 crc kubenswrapper[4685]: E1204 07:12:17.126219 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:12:29 crc kubenswrapper[4685]: I1204 07:12:29.126230 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:12:29 crc kubenswrapper[4685]: E1204 07:12:29.126950 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:12:42 crc kubenswrapper[4685]: I1204 07:12:42.125766 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:12:42 crc kubenswrapper[4685]: E1204 07:12:42.126708 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:12:56 crc kubenswrapper[4685]: I1204 07:12:56.339519 4685 generic.go:334] "Generic (PLEG): container finished" podID="faf74226-3731-4b3a-9479-2289c54ac84d" containerID="535c83d81f3e45b7c9306f497e71da4e81f887b25194eb5f092063495003af4a" exitCode=0 Dec 04 07:12:56 crc kubenswrapper[4685]: I1204 07:12:56.339557 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-7bgqz" event={"ID":"faf74226-3731-4b3a-9479-2289c54ac84d","Type":"ContainerDied","Data":"535c83d81f3e45b7c9306f497e71da4e81f887b25194eb5f092063495003af4a"} Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.126248 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:12:57 crc kubenswrapper[4685]: E1204 07:12:57.126622 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.463105 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.508608 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v225f/crc-debug-7bgqz"] Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.519000 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v225f/crc-debug-7bgqz"] Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.536345 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/faf74226-3731-4b3a-9479-2289c54ac84d-host\") pod \"faf74226-3731-4b3a-9479-2289c54ac84d\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.536532 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faf74226-3731-4b3a-9479-2289c54ac84d-host" (OuterVolumeSpecName: "host") pod "faf74226-3731-4b3a-9479-2289c54ac84d" (UID: "faf74226-3731-4b3a-9479-2289c54ac84d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.536758 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b54t\" (UniqueName: \"kubernetes.io/projected/faf74226-3731-4b3a-9479-2289c54ac84d-kube-api-access-5b54t\") pod \"faf74226-3731-4b3a-9479-2289c54ac84d\" (UID: \"faf74226-3731-4b3a-9479-2289c54ac84d\") " Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.537468 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/faf74226-3731-4b3a-9479-2289c54ac84d-host\") on node \"crc\" DevicePath \"\"" Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.543171 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faf74226-3731-4b3a-9479-2289c54ac84d-kube-api-access-5b54t" (OuterVolumeSpecName: "kube-api-access-5b54t") pod "faf74226-3731-4b3a-9479-2289c54ac84d" (UID: "faf74226-3731-4b3a-9479-2289c54ac84d"). InnerVolumeSpecName "kube-api-access-5b54t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:12:57 crc kubenswrapper[4685]: I1204 07:12:57.639769 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b54t\" (UniqueName: \"kubernetes.io/projected/faf74226-3731-4b3a-9479-2289c54ac84d-kube-api-access-5b54t\") on node \"crc\" DevicePath \"\"" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.361390 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="458539a6cfefed32fc10e589e11ed1dc7aff0f07914f57be42ec8d2001342f1d" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.361495 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7bgqz" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.709350 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v225f/crc-debug-jtjmw"] Dec 04 07:12:58 crc kubenswrapper[4685]: E1204 07:12:58.710161 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf74226-3731-4b3a-9479-2289c54ac84d" containerName="container-00" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.710176 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf74226-3731-4b3a-9479-2289c54ac84d" containerName="container-00" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.710476 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="faf74226-3731-4b3a-9479-2289c54ac84d" containerName="container-00" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.711294 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.761145 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7972811-6207-4b31-8749-519a41b6fac3-host\") pod \"crc-debug-jtjmw\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.761336 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmhvs\" (UniqueName: \"kubernetes.io/projected/b7972811-6207-4b31-8749-519a41b6fac3-kube-api-access-lmhvs\") pod \"crc-debug-jtjmw\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.863228 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7972811-6207-4b31-8749-519a41b6fac3-host\") pod \"crc-debug-jtjmw\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.863315 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmhvs\" (UniqueName: \"kubernetes.io/projected/b7972811-6207-4b31-8749-519a41b6fac3-kube-api-access-lmhvs\") pod \"crc-debug-jtjmw\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.863465 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7972811-6207-4b31-8749-519a41b6fac3-host\") pod \"crc-debug-jtjmw\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:58 crc kubenswrapper[4685]: I1204 07:12:58.890855 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmhvs\" (UniqueName: \"kubernetes.io/projected/b7972811-6207-4b31-8749-519a41b6fac3-kube-api-access-lmhvs\") pod \"crc-debug-jtjmw\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.030116 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.140547 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faf74226-3731-4b3a-9479-2289c54ac84d" path="/var/lib/kubelet/pods/faf74226-3731-4b3a-9479-2289c54ac84d/volumes" Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.370612 4685 generic.go:334] "Generic (PLEG): container finished" podID="b7972811-6207-4b31-8749-519a41b6fac3" containerID="7994613f7f4301d6a509851c97b16576e47cf6c236e975da7d258dde4d731d82" exitCode=0 Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.370662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-jtjmw" event={"ID":"b7972811-6207-4b31-8749-519a41b6fac3","Type":"ContainerDied","Data":"7994613f7f4301d6a509851c97b16576e47cf6c236e975da7d258dde4d731d82"} Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.370968 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-jtjmw" event={"ID":"b7972811-6207-4b31-8749-519a41b6fac3","Type":"ContainerStarted","Data":"26ad495c6cc27468988bb57aa5367acd03a1842d66015ff0b907700db72f7d78"} Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.883653 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v225f/crc-debug-jtjmw"] Dec 04 07:12:59 crc kubenswrapper[4685]: I1204 07:12:59.890610 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v225f/crc-debug-jtjmw"] Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.512379 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.591205 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7972811-6207-4b31-8749-519a41b6fac3-host\") pod \"b7972811-6207-4b31-8749-519a41b6fac3\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.591301 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmhvs\" (UniqueName: \"kubernetes.io/projected/b7972811-6207-4b31-8749-519a41b6fac3-kube-api-access-lmhvs\") pod \"b7972811-6207-4b31-8749-519a41b6fac3\" (UID: \"b7972811-6207-4b31-8749-519a41b6fac3\") " Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.591331 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7972811-6207-4b31-8749-519a41b6fac3-host" (OuterVolumeSpecName: "host") pod "b7972811-6207-4b31-8749-519a41b6fac3" (UID: "b7972811-6207-4b31-8749-519a41b6fac3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.591919 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7972811-6207-4b31-8749-519a41b6fac3-host\") on node \"crc\" DevicePath \"\"" Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.597705 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7972811-6207-4b31-8749-519a41b6fac3-kube-api-access-lmhvs" (OuterVolumeSpecName: "kube-api-access-lmhvs") pod "b7972811-6207-4b31-8749-519a41b6fac3" (UID: "b7972811-6207-4b31-8749-519a41b6fac3"). InnerVolumeSpecName "kube-api-access-lmhvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:13:00 crc kubenswrapper[4685]: I1204 07:13:00.693945 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmhvs\" (UniqueName: \"kubernetes.io/projected/b7972811-6207-4b31-8749-519a41b6fac3-kube-api-access-lmhvs\") on node \"crc\" DevicePath \"\"" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.076162 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v225f/crc-debug-7m7dr"] Dec 04 07:13:01 crc kubenswrapper[4685]: E1204 07:13:01.076560 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7972811-6207-4b31-8749-519a41b6fac3" containerName="container-00" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.076574 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7972811-6207-4b31-8749-519a41b6fac3" containerName="container-00" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.076763 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7972811-6207-4b31-8749-519a41b6fac3" containerName="container-00" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.077532 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.137018 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7972811-6207-4b31-8749-519a41b6fac3" path="/var/lib/kubelet/pods/b7972811-6207-4b31-8749-519a41b6fac3/volumes" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.203553 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95jbk\" (UniqueName: \"kubernetes.io/projected/19c4b1a8-6e2e-44ff-a331-20a49227769b-kube-api-access-95jbk\") pod \"crc-debug-7m7dr\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.203605 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19c4b1a8-6e2e-44ff-a331-20a49227769b-host\") pod \"crc-debug-7m7dr\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.305926 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95jbk\" (UniqueName: \"kubernetes.io/projected/19c4b1a8-6e2e-44ff-a331-20a49227769b-kube-api-access-95jbk\") pod \"crc-debug-7m7dr\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.306044 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19c4b1a8-6e2e-44ff-a331-20a49227769b-host\") pod \"crc-debug-7m7dr\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.306253 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19c4b1a8-6e2e-44ff-a331-20a49227769b-host\") pod \"crc-debug-7m7dr\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.338400 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95jbk\" (UniqueName: \"kubernetes.io/projected/19c4b1a8-6e2e-44ff-a331-20a49227769b-kube-api-access-95jbk\") pod \"crc-debug-7m7dr\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.391081 4685 scope.go:117] "RemoveContainer" containerID="7994613f7f4301d6a509851c97b16576e47cf6c236e975da7d258dde4d731d82" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.391142 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-jtjmw" Dec 04 07:13:01 crc kubenswrapper[4685]: I1204 07:13:01.410759 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:01 crc kubenswrapper[4685]: W1204 07:13:01.440888 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19c4b1a8_6e2e_44ff_a331_20a49227769b.slice/crio-666acf425a68890b77a2b7f155d00776be2fba1689d5322da5b8abc73434daf1 WatchSource:0}: Error finding container 666acf425a68890b77a2b7f155d00776be2fba1689d5322da5b8abc73434daf1: Status 404 returned error can't find the container with id 666acf425a68890b77a2b7f155d00776be2fba1689d5322da5b8abc73434daf1 Dec 04 07:13:02 crc kubenswrapper[4685]: I1204 07:13:02.407909 4685 generic.go:334] "Generic (PLEG): container finished" podID="19c4b1a8-6e2e-44ff-a331-20a49227769b" containerID="c9fa340462849c221098fa3d7ae4549fa7390a5ad3636898d9fe27a7c62692fe" exitCode=0 Dec 04 07:13:02 crc kubenswrapper[4685]: I1204 07:13:02.408071 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-7m7dr" event={"ID":"19c4b1a8-6e2e-44ff-a331-20a49227769b","Type":"ContainerDied","Data":"c9fa340462849c221098fa3d7ae4549fa7390a5ad3636898d9fe27a7c62692fe"} Dec 04 07:13:02 crc kubenswrapper[4685]: I1204 07:13:02.408894 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/crc-debug-7m7dr" event={"ID":"19c4b1a8-6e2e-44ff-a331-20a49227769b","Type":"ContainerStarted","Data":"666acf425a68890b77a2b7f155d00776be2fba1689d5322da5b8abc73434daf1"} Dec 04 07:13:02 crc kubenswrapper[4685]: I1204 07:13:02.472849 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v225f/crc-debug-7m7dr"] Dec 04 07:13:02 crc kubenswrapper[4685]: I1204 07:13:02.483976 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v225f/crc-debug-7m7dr"] Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.525294 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.653686 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19c4b1a8-6e2e-44ff-a331-20a49227769b-host\") pod \"19c4b1a8-6e2e-44ff-a331-20a49227769b\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.653824 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19c4b1a8-6e2e-44ff-a331-20a49227769b-host" (OuterVolumeSpecName: "host") pod "19c4b1a8-6e2e-44ff-a331-20a49227769b" (UID: "19c4b1a8-6e2e-44ff-a331-20a49227769b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.653882 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95jbk\" (UniqueName: \"kubernetes.io/projected/19c4b1a8-6e2e-44ff-a331-20a49227769b-kube-api-access-95jbk\") pod \"19c4b1a8-6e2e-44ff-a331-20a49227769b\" (UID: \"19c4b1a8-6e2e-44ff-a331-20a49227769b\") " Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.654395 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19c4b1a8-6e2e-44ff-a331-20a49227769b-host\") on node \"crc\" DevicePath \"\"" Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.662593 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19c4b1a8-6e2e-44ff-a331-20a49227769b-kube-api-access-95jbk" (OuterVolumeSpecName: "kube-api-access-95jbk") pod "19c4b1a8-6e2e-44ff-a331-20a49227769b" (UID: "19c4b1a8-6e2e-44ff-a331-20a49227769b"). InnerVolumeSpecName "kube-api-access-95jbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:13:03 crc kubenswrapper[4685]: I1204 07:13:03.756690 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95jbk\" (UniqueName: \"kubernetes.io/projected/19c4b1a8-6e2e-44ff-a331-20a49227769b-kube-api-access-95jbk\") on node \"crc\" DevicePath \"\"" Dec 04 07:13:04 crc kubenswrapper[4685]: I1204 07:13:04.428854 4685 scope.go:117] "RemoveContainer" containerID="c9fa340462849c221098fa3d7ae4549fa7390a5ad3636898d9fe27a7c62692fe" Dec 04 07:13:04 crc kubenswrapper[4685]: I1204 07:13:04.428884 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/crc-debug-7m7dr" Dec 04 07:13:05 crc kubenswrapper[4685]: I1204 07:13:05.143735 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19c4b1a8-6e2e-44ff-a331-20a49227769b" path="/var/lib/kubelet/pods/19c4b1a8-6e2e-44ff-a331-20a49227769b/volumes" Dec 04 07:13:10 crc kubenswrapper[4685]: I1204 07:13:10.125957 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:13:10 crc kubenswrapper[4685]: E1204 07:13:10.126825 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.259240 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-b6b84b88b-bjcmm_66becbc7-a80f-45d5-a711-f0c50c304034/barbican-api/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.363868 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-b6b84b88b-bjcmm_66becbc7-a80f-45d5-a711-f0c50c304034/barbican-api-log/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.475761 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64bd497db6-z9fft_1d7a178f-f4ba-4a18-ab97-f2d256e1edc1/barbican-keystone-listener/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.496125 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64bd497db6-z9fft_1d7a178f-f4ba-4a18-ab97-f2d256e1edc1/barbican-keystone-listener-log/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.658664 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84f9d55fbc-qkcbh_a1d37923-dce0-47f0-990f-92efc1754323/barbican-worker/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.681018 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84f9d55fbc-qkcbh_a1d37923-dce0-47f0-990f-92efc1754323/barbican-worker-log/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.798154 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2_398c2009-ba3b-49fa-b49d-838d80fe4bea/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.877535 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/ceilometer-central-agent/0.log" Dec 04 07:13:17 crc kubenswrapper[4685]: I1204 07:13:17.986824 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/ceilometer-notification-agent/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.021771 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/proxy-httpd/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.086392 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/sg-core/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.230941 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ad866e82-c1d3-4a38-953e-917ef323430b/cinder-api-log/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.262499 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ad866e82-c1d3-4a38-953e-917ef323430b/cinder-api/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.422444 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_67076e23-eac4-4e2a-944a-8687060343ad/cinder-scheduler/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.462082 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_67076e23-eac4-4e2a-944a-8687060343ad/probe/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.602009 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w_117d2695-776f-4596-9668-d99f7bd3f2bc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.720770 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng_2729f4b1-7245-40d2-a0e0-f4bb72e2fb05/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.838353 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-vknm9_22be434f-4bd4-4e59-b980-ae992dd4cf7d/init/0.log" Dec 04 07:13:18 crc kubenswrapper[4685]: I1204 07:13:18.991869 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-vknm9_22be434f-4bd4-4e59-b980-ae992dd4cf7d/init/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.047902 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-vknm9_22be434f-4bd4-4e59-b980-ae992dd4cf7d/dnsmasq-dns/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.107511 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-97x4w_701d7b10-8594-480e-8b64-ee2f6513b970/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.291982 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8c17edd9-e27f-4397-bfad-baeb4684038e/glance-httpd/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.326249 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8c17edd9-e27f-4397-bfad-baeb4684038e/glance-log/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.465343 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_55282f54-fc3a-418a-8eeb-3e4fa482b32b/glance-httpd/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.488901 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_55282f54-fc3a-418a-8eeb-3e4fa482b32b/glance-log/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.710381 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c9548d9dd-h5vqf_0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236/horizon/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.791855 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq_e75fdf41-c07d-4fdd-9abc-334e2164b96e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:19 crc kubenswrapper[4685]: I1204 07:13:19.980548 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c9548d9dd-h5vqf_0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236/horizon-log/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.009190 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-r22kq_d64ce345-0870-44f8-8cfb-da7b8cd6da34/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.308417 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413861-26zsj_f45f0571-fa49-4f6a-a36d-82dadc904495/keystone-cron/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.380999 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-679cdf976b-2nlz9_c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd/keystone-api/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.508922 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a504c958-f275-4b2f-84b7-566b0273f398/kube-state-metrics/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.545916 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c_5ae85205-468a-4d79-855f-0efa0c8f3534/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.885796 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-964564b4f-vkm6n_0452cb7f-8545-42b9-9456-b29aeaf9c391/neutron-httpd/0.log" Dec 04 07:13:20 crc kubenswrapper[4685]: I1204 07:13:20.962744 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-964564b4f-vkm6n_0452cb7f-8545-42b9-9456-b29aeaf9c391/neutron-api/0.log" Dec 04 07:13:21 crc kubenswrapper[4685]: I1204 07:13:21.099481 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq_b314b13b-4d0c-47f2-8fe8-1ab32ad74f28/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:21 crc kubenswrapper[4685]: I1204 07:13:21.467263 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_44873bdc-8a05-43b8-897e-1f44d10563f7/nova-cell0-conductor-conductor/0.log" Dec 04 07:13:21 crc kubenswrapper[4685]: I1204 07:13:21.570876 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_bd4268b9-da6d-496f-96e6-acf27d7e82fa/nova-api-log/0.log" Dec 04 07:13:21 crc kubenswrapper[4685]: I1204 07:13:21.733137 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_bd4268b9-da6d-496f-96e6-acf27d7e82fa/nova-api-api/0.log" Dec 04 07:13:21 crc kubenswrapper[4685]: I1204 07:13:21.773482 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2be52c9a-2015-4e17-a148-525b72b20ee0/nova-cell1-conductor-conductor/0.log" Dec 04 07:13:21 crc kubenswrapper[4685]: I1204 07:13:21.909970 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.053676 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hhdpc_1d8d290a-eebc-4783-be38-9b592fa66f25/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.219098 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_104b994d-e75e-4567-9877-f316b73024c4/nova-metadata-log/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.456927 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8bb3a776-b2f9-448c-b44e-c106669c7c19/nova-scheduler-scheduler/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.539120 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_536ebc06-2d82-419d-ac24-957fbc6734c9/mysql-bootstrap/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.756775 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_536ebc06-2d82-419d-ac24-957fbc6734c9/mysql-bootstrap/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.772326 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_536ebc06-2d82-419d-ac24-957fbc6734c9/galera/0.log" Dec 04 07:13:22 crc kubenswrapper[4685]: I1204 07:13:22.963277 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_577f79bb-8577-412a-92e1-bf56244becce/mysql-bootstrap/0.log" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.132859 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:13:23 crc kubenswrapper[4685]: E1204 07:13:23.133362 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.236603 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_577f79bb-8577-412a-92e1-bf56244becce/mysql-bootstrap/0.log" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.261013 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_577f79bb-8577-412a-92e1-bf56244becce/galera/0.log" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.452991 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_104b994d-e75e-4567-9877-f316b73024c4/nova-metadata-metadata/0.log" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.458977 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6a4a5bce-68e1-448c-a00c-d7fc97e69566/openstackclient/0.log" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.501825 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7njcr_29b767dd-48fe-49d1-969a-e0347cf9a2c4/openstack-network-exporter/0.log" Dec 04 07:13:23 crc kubenswrapper[4685]: I1204 07:13:23.694854 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovsdb-server-init/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.007042 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovsdb-server/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.070234 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovsdb-server-init/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.120880 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovs-vswitchd/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.259581 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z562k_2db0269f-8ebe-48f0-81de-081da72da040/ovn-controller/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.406265 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7nn2f_252256a3-494e-45de-ae52-6f77715b468e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.452861 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_53bf9890-b3a9-4102-9b2f-f7711e7dbe70/openstack-network-exporter/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.559214 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_53bf9890-b3a9-4102-9b2f-f7711e7dbe70/ovn-northd/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.799282 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_78fd03f6-d85a-4348-a910-2cd71864b1fc/openstack-network-exporter/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.853964 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_78fd03f6-d85a-4348-a910-2cd71864b1fc/ovsdbserver-nb/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.969063 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c99c06e8-97e8-44d4-b532-a2ca943d826a/openstack-network-exporter/0.log" Dec 04 07:13:24 crc kubenswrapper[4685]: I1204 07:13:24.997084 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c99c06e8-97e8-44d4-b532-a2ca943d826a/ovsdbserver-sb/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.235755 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6bc7689ffb-9thrj_4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55/placement-log/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.318953 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6bc7689ffb-9thrj_4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55/placement-api/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.329817 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b/setup-container/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.580399 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b/setup-container/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.623573 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b87c3420-10be-48df-89b4-c388608be613/setup-container/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.676853 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b/rabbitmq/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.831296 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b87c3420-10be-48df-89b4-c388608be613/setup-container/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.892222 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b87c3420-10be-48df-89b4-c388608be613/rabbitmq/0.log" Dec 04 07:13:25 crc kubenswrapper[4685]: I1204 07:13:25.936811 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp_f64013bb-a12d-4cfd-93de-741c83048e77/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.058234 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tx98n_08644b36-c6a1-4c4f-9e4d-89e861f2c197/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.140105 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd_7d609a36-8df9-42ab-8160-c6f8b49880d9/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.376952 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rxvkk_d20c1b8a-7db1-4d00-8907-88a6a3e3d312/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.436823 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-xldx9_f33640dd-287a-4de5-be7a-bc4fa6fd72fe/ssh-known-hosts-edpm-deployment/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.688447 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-69f76768ff-42jpz_70d30e39-7d08-4587-90cb-48fc2c868dbc/proxy-server/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.715514 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-69f76768ff-42jpz_70d30e39-7d08-4587-90cb-48fc2c868dbc/proxy-httpd/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.781945 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9bq7c_996bb757-370f-4635-9b6f-923ecd70a8ee/swift-ring-rebalance/0.log" Dec 04 07:13:26 crc kubenswrapper[4685]: I1204 07:13:26.957849 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-auditor/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.024325 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-reaper/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.107670 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-replicator/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.181429 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-auditor/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.238068 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-server/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.259590 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-replicator/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.335068 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-server/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.391977 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-updater/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.468800 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-auditor/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.503665 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-expirer/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.559130 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-replicator/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.582030 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-server/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.683671 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-updater/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.703720 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/rsync/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.789518 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/swift-recon-cron/0.log" Dec 04 07:13:27 crc kubenswrapper[4685]: I1204 07:13:27.990331 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4_175a4186-effa-431c-9808-33f3fa73119b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:28 crc kubenswrapper[4685]: I1204 07:13:28.079389 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c8fff46b-61e1-48da-a047-9ace06e11177/tempest-tests-tempest-tests-runner/0.log" Dec 04 07:13:28 crc kubenswrapper[4685]: I1204 07:13:28.187338 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_98d35031-8372-42b2-9b45-e068ff8b4c08/test-operator-logs-container/0.log" Dec 04 07:13:28 crc kubenswrapper[4685]: I1204 07:13:28.318597 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd_080fc1a1-8d27-4b03-bd8f-04d200f62da2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:13:34 crc kubenswrapper[4685]: I1204 07:13:34.125883 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:13:34 crc kubenswrapper[4685]: E1204 07:13:34.126568 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:13:38 crc kubenswrapper[4685]: I1204 07:13:38.785776 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7/memcached/0.log" Dec 04 07:13:45 crc kubenswrapper[4685]: I1204 07:13:45.126561 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:13:45 crc kubenswrapper[4685]: E1204 07:13:45.127458 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:13:54 crc kubenswrapper[4685]: I1204 07:13:54.638386 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/util/0.log" Dec 04 07:13:54 crc kubenswrapper[4685]: I1204 07:13:54.828483 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/util/0.log" Dec 04 07:13:54 crc kubenswrapper[4685]: I1204 07:13:54.833154 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/pull/0.log" Dec 04 07:13:54 crc kubenswrapper[4685]: I1204 07:13:54.846752 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/pull/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.013749 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/util/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.014319 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/pull/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.019306 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/extract/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.205088 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-h8fbl_c9f1a1e0-3fd0-4c69-8475-9ed056330d57/kube-rbac-proxy/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.287401 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-s8zzd_4c3ba289-3038-4c41-881a-d53e4f54bc47/kube-rbac-proxy/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.292870 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-h8fbl_c9f1a1e0-3fd0-4c69-8475-9ed056330d57/manager/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.429129 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-s8zzd_4c3ba289-3038-4c41-881a-d53e4f54bc47/manager/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.490377 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6pgsq_7aa55d54-0ab9-4b25-8226-095d3703ca03/manager/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.525063 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6pgsq_7aa55d54-0ab9-4b25-8226-095d3703ca03/kube-rbac-proxy/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.615877 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-czngd_231797d5-080c-4e7f-ad27-186a01bf4094/kube-rbac-proxy/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.780341 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-czngd_231797d5-080c-4e7f-ad27-186a01bf4094/manager/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.863806 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rt2mx_87b87b6f-538b-4ba3-ba4c-994924ffe76e/kube-rbac-proxy/0.log" Dec 04 07:13:55 crc kubenswrapper[4685]: I1204 07:13:55.891319 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rt2mx_87b87b6f-538b-4ba3-ba4c-994924ffe76e/manager/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.004608 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-wd5jl_211501da-f7e5-405e-b1cb-2673ca2b82c1/kube-rbac-proxy/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.069153 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-wd5jl_211501da-f7e5-405e-b1cb-2673ca2b82c1/manager/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.127018 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:13:56 crc kubenswrapper[4685]: E1204 07:13:56.127371 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.160118 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-8fjlw_d55943b8-039d-4747-b341-d0fe2e0f6df7/kube-rbac-proxy/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.339591 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-8fjlw_d55943b8-039d-4747-b341-d0fe2e0f6df7/manager/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.361691 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qqd9p_ee4e7fd4-46c8-4aac-be76-5097d9f76ac2/kube-rbac-proxy/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.371428 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qqd9p_ee4e7fd4-46c8-4aac-be76-5097d9f76ac2/manager/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.502206 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-xql6q_cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1/kube-rbac-proxy/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.655122 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-xql6q_cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1/manager/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.665910 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rrk46_4af99755-81d3-43ec-a924-a3dcc4347315/kube-rbac-proxy/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.717112 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rrk46_4af99755-81d3-43ec-a924-a3dcc4347315/manager/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.871382 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-dnkjj_89fc2488-fa6d-4515-923e-115aa45e23d4/kube-rbac-proxy/0.log" Dec 04 07:13:56 crc kubenswrapper[4685]: I1204 07:13:56.885559 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-dnkjj_89fc2488-fa6d-4515-923e-115aa45e23d4/manager/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.009295 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7t9xh_e9b56e59-1bc8-4608-bdfe-2901e30c7009/kube-rbac-proxy/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.130587 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7t9xh_e9b56e59-1bc8-4608-bdfe-2901e30c7009/manager/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.218663 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jlm87_c7545a63-4651-4d38-a251-9de38ce5a226/kube-rbac-proxy/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.291334 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jlm87_c7545a63-4651-4d38-a251-9de38ce5a226/manager/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.526997 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-v2lxd_09248202-ba20-4b70-95ab-1a7e67c04ef7/kube-rbac-proxy/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.586434 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-v2lxd_09248202-ba20-4b70-95ab-1a7e67c04ef7/manager/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.726791 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9_3ed35c5e-42ad-433c-a044-0fdac4c8ad06/manager/0.log" Dec 04 07:13:57 crc kubenswrapper[4685]: I1204 07:13:57.746247 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9_3ed35c5e-42ad-433c-a044-0fdac4c8ad06/kube-rbac-proxy/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.206761 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-7m62c_569ea63b-9c26-411f-a899-6b881743135f/registry-server/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.234316 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-778b7fc6d-66vch_90f5cdd3-8605-4cb5-9aea-83d57954bc17/operator/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.293860 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-cdbkv_efab633f-9698-443f-8175-279a82810d4d/kube-rbac-proxy/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.478714 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-cdbkv_efab633f-9698-443f-8175-279a82810d4d/manager/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.519742 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nhb96_e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13/kube-rbac-proxy/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.610870 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nhb96_e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13/manager/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.703068 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-tmmg4_ecb2e435-29cf-4b37-a0bb-4f2441cab009/operator/0.log" Dec 04 07:13:58 crc kubenswrapper[4685]: I1204 07:13:58.922560 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xchfb_18233d1f-7861-4bd5-8643-f464fce185ca/kube-rbac-proxy/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.036012 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-c64mf_9cee5367-7803-44a8-b047-66e4311f9e0a/kube-rbac-proxy/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.039225 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xchfb_18233d1f-7861-4bd5-8643-f464fce185ca/manager/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.201285 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54986f46d6-f8k5j_541675b0-83d5-49cf-8209-2de2d8385f3b/manager/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.226012 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-c64mf_9cee5367-7803-44a8-b047-66e4311f9e0a/manager/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.262801 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7wr8_e3ac8015-c6a6-4837-9118-0de5f1b351df/kube-rbac-proxy/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.351145 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7wr8_e3ac8015-c6a6-4837-9118-0de5f1b351df/manager/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.434656 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-zzn46_639879a9-dc01-4145-916b-8d568130af06/kube-rbac-proxy/0.log" Dec 04 07:13:59 crc kubenswrapper[4685]: I1204 07:13:59.435068 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-zzn46_639879a9-dc01-4145-916b-8d568130af06/manager/0.log" Dec 04 07:14:09 crc kubenswrapper[4685]: I1204 07:14:09.125769 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:14:09 crc kubenswrapper[4685]: E1204 07:14:09.126724 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:14:17 crc kubenswrapper[4685]: I1204 07:14:17.065575 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2wd6d_007053aa-6733-47ec-aeb6-7b6753d58c15/control-plane-machine-set-operator/0.log" Dec 04 07:14:17 crc kubenswrapper[4685]: I1204 07:14:17.138941 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b9hw7_25743472-93a4-4d4e-bf54-1d5cef4025a2/kube-rbac-proxy/0.log" Dec 04 07:14:17 crc kubenswrapper[4685]: I1204 07:14:17.173434 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b9hw7_25743472-93a4-4d4e-bf54-1d5cef4025a2/machine-api-operator/0.log" Dec 04 07:14:21 crc kubenswrapper[4685]: I1204 07:14:21.126140 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:14:21 crc kubenswrapper[4685]: E1204 07:14:21.126840 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:14:29 crc kubenswrapper[4685]: I1204 07:14:29.273155 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nqsrd_8ffffd54-b5c5-434a-b2d3-6afb6f12a254/cert-manager-controller/0.log" Dec 04 07:14:29 crc kubenswrapper[4685]: I1204 07:14:29.461861 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x4tmf_49a6e18b-01a4-44b9-8487-368f096cf059/cert-manager-cainjector/0.log" Dec 04 07:14:29 crc kubenswrapper[4685]: I1204 07:14:29.521628 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-28r8x_a33af606-6a04-4d1b-a5e5-b709163d4d89/cert-manager-webhook/0.log" Dec 04 07:14:33 crc kubenswrapper[4685]: I1204 07:14:33.133602 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:14:33 crc kubenswrapper[4685]: E1204 07:14:33.134670 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:14:41 crc kubenswrapper[4685]: I1204 07:14:41.765075 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-rnj2j_f77ee3f7-2978-4a54-ac7e-11a038bd3f5e/nmstate-console-plugin/0.log" Dec 04 07:14:41 crc kubenswrapper[4685]: I1204 07:14:41.950694 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-t55gj_bef52eb9-45f0-43e5-b675-e4d3e819a3ec/kube-rbac-proxy/0.log" Dec 04 07:14:41 crc kubenswrapper[4685]: I1204 07:14:41.965722 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wpz9x_a3048137-1da6-4838-99bf-f4ab08bf895c/nmstate-handler/0.log" Dec 04 07:14:42 crc kubenswrapper[4685]: I1204 07:14:42.027584 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-t55gj_bef52eb9-45f0-43e5-b675-e4d3e819a3ec/nmstate-metrics/0.log" Dec 04 07:14:42 crc kubenswrapper[4685]: I1204 07:14:42.152136 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bz488_f0b0d865-5ec1-4dff-ada5-658bd9676d32/nmstate-operator/0.log" Dec 04 07:14:42 crc kubenswrapper[4685]: I1204 07:14:42.241827 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-9hhls_ec56bcce-1006-4645-a67d-2a4405768a11/nmstate-webhook/0.log" Dec 04 07:14:46 crc kubenswrapper[4685]: I1204 07:14:46.126527 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:14:46 crc kubenswrapper[4685]: E1204 07:14:46.127058 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.372159 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d7jj8_62ade17b-938c-44f3-9af9-b5138439e87e/kube-rbac-proxy/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.413189 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d7jj8_62ade17b-938c-44f3-9af9-b5138439e87e/controller/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.544484 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.717334 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.728495 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.744370 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.761893 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.879586 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.904310 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.920202 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:14:56 crc kubenswrapper[4685]: I1204 07:14:56.928459 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.127387 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.138399 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.148266 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.187816 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/controller/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.322500 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/kube-rbac-proxy/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.346466 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/frr-metrics/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.419377 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/kube-rbac-proxy-frr/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.615737 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/reloader/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.656961 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-hcsrv_4eb6708c-7ef6-477d-a404-fd591df256b5/frr-k8s-webhook-server/0.log" Dec 04 07:14:57 crc kubenswrapper[4685]: I1204 07:14:57.905530 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57b788c8f7-tgv57_0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7/manager/0.log" Dec 04 07:14:58 crc kubenswrapper[4685]: I1204 07:14:58.068140 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86896cd6d4-h6dq8_c5b1e90c-cba0-40f0-b18a-4b7c8c52900c/webhook-server/0.log" Dec 04 07:14:58 crc kubenswrapper[4685]: I1204 07:14:58.226320 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5mm56_7ebc5585-22bd-459e-bd2d-db1bc027a03f/kube-rbac-proxy/0.log" Dec 04 07:14:58 crc kubenswrapper[4685]: I1204 07:14:58.766623 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5mm56_7ebc5585-22bd-459e-bd2d-db1bc027a03f/speaker/0.log" Dec 04 07:14:58 crc kubenswrapper[4685]: I1204 07:14:58.913096 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/frr/0.log" Dec 04 07:14:59 crc kubenswrapper[4685]: I1204 07:14:59.125370 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:14:59 crc kubenswrapper[4685]: E1204 07:14:59.125676 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.190843 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf"] Dec 04 07:15:00 crc kubenswrapper[4685]: E1204 07:15:00.191272 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19c4b1a8-6e2e-44ff-a331-20a49227769b" containerName="container-00" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.191285 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="19c4b1a8-6e2e-44ff-a331-20a49227769b" containerName="container-00" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.191485 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="19c4b1a8-6e2e-44ff-a331-20a49227769b" containerName="container-00" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.192091 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.195088 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.195088 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.210595 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf"] Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.355483 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j72q2\" (UniqueName: \"kubernetes.io/projected/df6d9f39-08c6-4a53-8f20-9e6a56091bba-kube-api-access-j72q2\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.355576 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df6d9f39-08c6-4a53-8f20-9e6a56091bba-secret-volume\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.355617 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df6d9f39-08c6-4a53-8f20-9e6a56091bba-config-volume\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.457370 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j72q2\" (UniqueName: \"kubernetes.io/projected/df6d9f39-08c6-4a53-8f20-9e6a56091bba-kube-api-access-j72q2\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.457462 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df6d9f39-08c6-4a53-8f20-9e6a56091bba-secret-volume\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.457499 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df6d9f39-08c6-4a53-8f20-9e6a56091bba-config-volume\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.458341 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df6d9f39-08c6-4a53-8f20-9e6a56091bba-config-volume\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.464967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df6d9f39-08c6-4a53-8f20-9e6a56091bba-secret-volume\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.477463 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j72q2\" (UniqueName: \"kubernetes.io/projected/df6d9f39-08c6-4a53-8f20-9e6a56091bba-kube-api-access-j72q2\") pod \"collect-profiles-29413875-fbrjf\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:00 crc kubenswrapper[4685]: I1204 07:15:00.513940 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:01 crc kubenswrapper[4685]: I1204 07:15:01.011822 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf"] Dec 04 07:15:01 crc kubenswrapper[4685]: I1204 07:15:01.551060 4685 generic.go:334] "Generic (PLEG): container finished" podID="df6d9f39-08c6-4a53-8f20-9e6a56091bba" containerID="d2d4a9daa65984a2a068f7e89705db403f83bb9a36c624fe833ef7c750059490" exitCode=0 Dec 04 07:15:01 crc kubenswrapper[4685]: I1204 07:15:01.551175 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" event={"ID":"df6d9f39-08c6-4a53-8f20-9e6a56091bba","Type":"ContainerDied","Data":"d2d4a9daa65984a2a068f7e89705db403f83bb9a36c624fe833ef7c750059490"} Dec 04 07:15:01 crc kubenswrapper[4685]: I1204 07:15:01.551489 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" event={"ID":"df6d9f39-08c6-4a53-8f20-9e6a56091bba","Type":"ContainerStarted","Data":"0bdc0a61f129a7bb36e641de7b729c8001ef6d43d5a458119ed3f115b00527d9"} Dec 04 07:15:02 crc kubenswrapper[4685]: I1204 07:15:02.944813 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.104524 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j72q2\" (UniqueName: \"kubernetes.io/projected/df6d9f39-08c6-4a53-8f20-9e6a56091bba-kube-api-access-j72q2\") pod \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.104637 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df6d9f39-08c6-4a53-8f20-9e6a56091bba-config-volume\") pod \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.104686 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df6d9f39-08c6-4a53-8f20-9e6a56091bba-secret-volume\") pod \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\" (UID: \"df6d9f39-08c6-4a53-8f20-9e6a56091bba\") " Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.105015 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6d9f39-08c6-4a53-8f20-9e6a56091bba-config-volume" (OuterVolumeSpecName: "config-volume") pod "df6d9f39-08c6-4a53-8f20-9e6a56091bba" (UID: "df6d9f39-08c6-4a53-8f20-9e6a56091bba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.105146 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df6d9f39-08c6-4a53-8f20-9e6a56091bba-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.110554 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6d9f39-08c6-4a53-8f20-9e6a56091bba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df6d9f39-08c6-4a53-8f20-9e6a56091bba" (UID: "df6d9f39-08c6-4a53-8f20-9e6a56091bba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.113544 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6d9f39-08c6-4a53-8f20-9e6a56091bba-kube-api-access-j72q2" (OuterVolumeSpecName: "kube-api-access-j72q2") pod "df6d9f39-08c6-4a53-8f20-9e6a56091bba" (UID: "df6d9f39-08c6-4a53-8f20-9e6a56091bba"). InnerVolumeSpecName "kube-api-access-j72q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.207097 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j72q2\" (UniqueName: \"kubernetes.io/projected/df6d9f39-08c6-4a53-8f20-9e6a56091bba-kube-api-access-j72q2\") on node \"crc\" DevicePath \"\"" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.207139 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df6d9f39-08c6-4a53-8f20-9e6a56091bba-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.570942 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" event={"ID":"df6d9f39-08c6-4a53-8f20-9e6a56091bba","Type":"ContainerDied","Data":"0bdc0a61f129a7bb36e641de7b729c8001ef6d43d5a458119ed3f115b00527d9"} Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.571012 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bdc0a61f129a7bb36e641de7b729c8001ef6d43d5a458119ed3f115b00527d9" Dec 04 07:15:03 crc kubenswrapper[4685]: I1204 07:15:03.571108 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413875-fbrjf" Dec 04 07:15:04 crc kubenswrapper[4685]: I1204 07:15:04.039425 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4"] Dec 04 07:15:04 crc kubenswrapper[4685]: I1204 07:15:04.047253 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413830-kqvz4"] Dec 04 07:15:05 crc kubenswrapper[4685]: I1204 07:15:05.139835 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8daa431f-999a-4213-a745-9d8b03aa028e" path="/var/lib/kubelet/pods/8daa431f-999a-4213-a745-9d8b03aa028e/volumes" Dec 04 07:15:10 crc kubenswrapper[4685]: I1204 07:15:10.126489 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:15:10 crc kubenswrapper[4685]: E1204 07:15:10.127593 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.205452 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/util/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.329143 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/util/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.406474 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/pull/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.433653 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/pull/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.584208 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/pull/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.604857 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/extract/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.638166 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/util/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.760731 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/util/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.956375 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/pull/0.log" Dec 04 07:15:12 crc kubenswrapper[4685]: I1204 07:15:12.971515 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/util/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.001037 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/pull/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.159533 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/util/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.180405 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/extract/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.180678 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/pull/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.483322 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-utilities/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.669466 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-utilities/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.693875 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-content/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.700905 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-content/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.880928 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-utilities/0.log" Dec 04 07:15:13 crc kubenswrapper[4685]: I1204 07:15:13.888474 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-content/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.167997 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-utilities/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.377298 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-content/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.427576 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-utilities/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.447241 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/registry-server/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.500211 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-content/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.701063 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-utilities/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.735367 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-content/0.log" Dec 04 07:15:14 crc kubenswrapper[4685]: I1204 07:15:14.969090 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-49wxx_c54a0f24-1600-4e6a-8c7c-00f003250e59/marketplace-operator/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.010939 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-utilities/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.381960 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/registry-server/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.403060 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-utilities/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.417081 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-content/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.457492 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-content/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.606785 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-utilities/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.631471 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-content/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.788002 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/registry-server/0.log" Dec 04 07:15:15 crc kubenswrapper[4685]: I1204 07:15:15.872214 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-utilities/0.log" Dec 04 07:15:16 crc kubenswrapper[4685]: I1204 07:15:16.018693 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-utilities/0.log" Dec 04 07:15:16 crc kubenswrapper[4685]: I1204 07:15:16.050551 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-content/0.log" Dec 04 07:15:16 crc kubenswrapper[4685]: I1204 07:15:16.051876 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-content/0.log" Dec 04 07:15:16 crc kubenswrapper[4685]: I1204 07:15:16.276314 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-utilities/0.log" Dec 04 07:15:16 crc kubenswrapper[4685]: I1204 07:15:16.284150 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-content/0.log" Dec 04 07:15:16 crc kubenswrapper[4685]: I1204 07:15:16.850881 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/registry-server/0.log" Dec 04 07:15:19 crc kubenswrapper[4685]: I1204 07:15:19.732749 4685 scope.go:117] "RemoveContainer" containerID="55c52dcec029a86895ba594f836748f637c993c60370e69352ae95b6419ec4ef" Dec 04 07:15:23 crc kubenswrapper[4685]: I1204 07:15:23.132143 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:15:23 crc kubenswrapper[4685]: E1204 07:15:23.136003 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:15:35 crc kubenswrapper[4685]: I1204 07:15:35.126066 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:15:35 crc kubenswrapper[4685]: E1204 07:15:35.126886 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:15:49 crc kubenswrapper[4685]: I1204 07:15:49.126542 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:15:49 crc kubenswrapper[4685]: E1204 07:15:49.127338 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:16:01 crc kubenswrapper[4685]: I1204 07:16:01.126486 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:16:01 crc kubenswrapper[4685]: E1204 07:16:01.127391 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:16:12 crc kubenswrapper[4685]: I1204 07:16:12.125938 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:16:13 crc kubenswrapper[4685]: I1204 07:16:13.226178 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"b50cf90d1b5e54d17b9b4973cea9196127ec31bda986876081b21e993142a848"} Dec 04 07:16:58 crc kubenswrapper[4685]: I1204 07:16:58.669225 4685 generic.go:334] "Generic (PLEG): container finished" podID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerID="6e305812b42877dd9791806a0cb6987062bf480adc82261425a4459316de1574" exitCode=0 Dec 04 07:16:58 crc kubenswrapper[4685]: I1204 07:16:58.669307 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v225f/must-gather-rtvxw" event={"ID":"7b59e18d-45b2-43e3-bdac-f28ac18243c6","Type":"ContainerDied","Data":"6e305812b42877dd9791806a0cb6987062bf480adc82261425a4459316de1574"} Dec 04 07:16:58 crc kubenswrapper[4685]: I1204 07:16:58.670742 4685 scope.go:117] "RemoveContainer" containerID="6e305812b42877dd9791806a0cb6987062bf480adc82261425a4459316de1574" Dec 04 07:16:59 crc kubenswrapper[4685]: I1204 07:16:59.687929 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v225f_must-gather-rtvxw_7b59e18d-45b2-43e3-bdac-f28ac18243c6/gather/0.log" Dec 04 07:17:08 crc kubenswrapper[4685]: I1204 07:17:08.103573 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v225f/must-gather-rtvxw"] Dec 04 07:17:08 crc kubenswrapper[4685]: I1204 07:17:08.104258 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-v225f/must-gather-rtvxw" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="copy" containerID="cri-o://53f81f0bc8f98fd95e9961b588f4e6e785e7c13410d3ff5c1a20cb309d72a38b" gracePeriod=2 Dec 04 07:17:08 crc kubenswrapper[4685]: I1204 07:17:08.114156 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v225f/must-gather-rtvxw"] Dec 04 07:17:08 crc kubenswrapper[4685]: I1204 07:17:08.766499 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v225f_must-gather-rtvxw_7b59e18d-45b2-43e3-bdac-f28ac18243c6/copy/0.log" Dec 04 07:17:08 crc kubenswrapper[4685]: I1204 07:17:08.767166 4685 generic.go:334] "Generic (PLEG): container finished" podID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerID="53f81f0bc8f98fd95e9961b588f4e6e785e7c13410d3ff5c1a20cb309d72a38b" exitCode=143 Dec 04 07:17:08 crc kubenswrapper[4685]: I1204 07:17:08.767212 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d37659c0e9d7ed9d2a4b1620b77fbda1cc1f3b6fb9cb0a1a63496364c934740b" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.155342 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v225f_must-gather-rtvxw_7b59e18d-45b2-43e3-bdac-f28ac18243c6/copy/0.log" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.155809 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.267248 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tglfm\" (UniqueName: \"kubernetes.io/projected/7b59e18d-45b2-43e3-bdac-f28ac18243c6-kube-api-access-tglfm\") pod \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.267439 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7b59e18d-45b2-43e3-bdac-f28ac18243c6-must-gather-output\") pod \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\" (UID: \"7b59e18d-45b2-43e3-bdac-f28ac18243c6\") " Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.273195 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b59e18d-45b2-43e3-bdac-f28ac18243c6-kube-api-access-tglfm" (OuterVolumeSpecName: "kube-api-access-tglfm") pod "7b59e18d-45b2-43e3-bdac-f28ac18243c6" (UID: "7b59e18d-45b2-43e3-bdac-f28ac18243c6"). InnerVolumeSpecName "kube-api-access-tglfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.370208 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tglfm\" (UniqueName: \"kubernetes.io/projected/7b59e18d-45b2-43e3-bdac-f28ac18243c6-kube-api-access-tglfm\") on node \"crc\" DevicePath \"\"" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.403844 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b59e18d-45b2-43e3-bdac-f28ac18243c6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7b59e18d-45b2-43e3-bdac-f28ac18243c6" (UID: "7b59e18d-45b2-43e3-bdac-f28ac18243c6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.472585 4685 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7b59e18d-45b2-43e3-bdac-f28ac18243c6-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 07:17:09 crc kubenswrapper[4685]: I1204 07:17:09.778777 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v225f/must-gather-rtvxw" Dec 04 07:17:11 crc kubenswrapper[4685]: I1204 07:17:11.138885 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" path="/var/lib/kubelet/pods/7b59e18d-45b2-43e3-bdac-f28ac18243c6/volumes" Dec 04 07:18:19 crc kubenswrapper[4685]: I1204 07:18:19.850517 4685 scope.go:117] "RemoveContainer" containerID="53f81f0bc8f98fd95e9961b588f4e6e785e7c13410d3ff5c1a20cb309d72a38b" Dec 04 07:18:19 crc kubenswrapper[4685]: I1204 07:18:19.892843 4685 scope.go:117] "RemoveContainer" containerID="6e305812b42877dd9791806a0cb6987062bf480adc82261425a4459316de1574" Dec 04 07:18:19 crc kubenswrapper[4685]: I1204 07:18:19.928878 4685 scope.go:117] "RemoveContainer" containerID="535c83d81f3e45b7c9306f497e71da4e81f887b25194eb5f092063495003af4a" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.257630 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vbng9"] Dec 04 07:18:30 crc kubenswrapper[4685]: E1204 07:18:30.259106 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df6d9f39-08c6-4a53-8f20-9e6a56091bba" containerName="collect-profiles" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.259125 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="df6d9f39-08c6-4a53-8f20-9e6a56091bba" containerName="collect-profiles" Dec 04 07:18:30 crc kubenswrapper[4685]: E1204 07:18:30.259167 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="copy" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.259176 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="copy" Dec 04 07:18:30 crc kubenswrapper[4685]: E1204 07:18:30.259204 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="gather" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.259213 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="gather" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.259698 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="df6d9f39-08c6-4a53-8f20-9e6a56091bba" containerName="collect-profiles" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.259726 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="gather" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.259754 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b59e18d-45b2-43e3-bdac-f28ac18243c6" containerName="copy" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.264706 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.286082 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbng9"] Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.386712 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-utilities\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.386923 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqhzj\" (UniqueName: \"kubernetes.io/projected/61684126-28a3-458b-92ad-4500aae5fc0e-kube-api-access-cqhzj\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.387086 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-catalog-content\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.488933 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-utilities\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.489268 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqhzj\" (UniqueName: \"kubernetes.io/projected/61684126-28a3-458b-92ad-4500aae5fc0e-kube-api-access-cqhzj\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.489318 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-catalog-content\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.489431 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-utilities\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.489686 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-catalog-content\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.513533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqhzj\" (UniqueName: \"kubernetes.io/projected/61684126-28a3-458b-92ad-4500aae5fc0e-kube-api-access-cqhzj\") pod \"redhat-marketplace-vbng9\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:30 crc kubenswrapper[4685]: I1204 07:18:30.611757 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:31 crc kubenswrapper[4685]: I1204 07:18:31.071102 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbng9"] Dec 04 07:18:31 crc kubenswrapper[4685]: W1204 07:18:31.078029 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61684126_28a3_458b_92ad_4500aae5fc0e.slice/crio-9bfa2df458d29f2194056845ba16078e133baa8141a4dcadc5b630c2208613b1 WatchSource:0}: Error finding container 9bfa2df458d29f2194056845ba16078e133baa8141a4dcadc5b630c2208613b1: Status 404 returned error can't find the container with id 9bfa2df458d29f2194056845ba16078e133baa8141a4dcadc5b630c2208613b1 Dec 04 07:18:31 crc kubenswrapper[4685]: I1204 07:18:31.606530 4685 generic.go:334] "Generic (PLEG): container finished" podID="61684126-28a3-458b-92ad-4500aae5fc0e" containerID="df580f352cba18b0a4ecaaf1d4ad9fe60ea58bb6a0200a4a7ab0b9ff4562091a" exitCode=0 Dec 04 07:18:31 crc kubenswrapper[4685]: I1204 07:18:31.606611 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbng9" event={"ID":"61684126-28a3-458b-92ad-4500aae5fc0e","Type":"ContainerDied","Data":"df580f352cba18b0a4ecaaf1d4ad9fe60ea58bb6a0200a4a7ab0b9ff4562091a"} Dec 04 07:18:31 crc kubenswrapper[4685]: I1204 07:18:31.606908 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbng9" event={"ID":"61684126-28a3-458b-92ad-4500aae5fc0e","Type":"ContainerStarted","Data":"9bfa2df458d29f2194056845ba16078e133baa8141a4dcadc5b630c2208613b1"} Dec 04 07:18:31 crc kubenswrapper[4685]: I1204 07:18:31.609466 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.630661 4685 generic.go:334] "Generic (PLEG): container finished" podID="61684126-28a3-458b-92ad-4500aae5fc0e" containerID="534a8a85d55c9f6c46952d32588fd1d1a43aa6de9a3d672d25b6607cf2a8580a" exitCode=0 Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.630780 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbng9" event={"ID":"61684126-28a3-458b-92ad-4500aae5fc0e","Type":"ContainerDied","Data":"534a8a85d55c9f6c46952d32588fd1d1a43aa6de9a3d672d25b6607cf2a8580a"} Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.634839 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8hxqf"] Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.638453 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.664540 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8hxqf"] Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.733599 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-utilities\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.733934 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flmg2\" (UniqueName: \"kubernetes.io/projected/252c599c-8864-414d-a940-467c7fafa274-kube-api-access-flmg2\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.734379 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-catalog-content\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.836196 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-utilities\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.836258 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flmg2\" (UniqueName: \"kubernetes.io/projected/252c599c-8864-414d-a940-467c7fafa274-kube-api-access-flmg2\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.836319 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-catalog-content\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.836951 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-catalog-content\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.837259 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-utilities\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:32 crc kubenswrapper[4685]: I1204 07:18:32.867760 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flmg2\" (UniqueName: \"kubernetes.io/projected/252c599c-8864-414d-a940-467c7fafa274-kube-api-access-flmg2\") pod \"certified-operators-8hxqf\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:33 crc kubenswrapper[4685]: I1204 07:18:33.008892 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:33 crc kubenswrapper[4685]: I1204 07:18:33.513037 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8hxqf"] Dec 04 07:18:33 crc kubenswrapper[4685]: I1204 07:18:33.640782 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerStarted","Data":"27c6b9132afcce707782080aeaf7358d92c8b9499f939fe23a04be68d559a896"} Dec 04 07:18:33 crc kubenswrapper[4685]: I1204 07:18:33.643215 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbng9" event={"ID":"61684126-28a3-458b-92ad-4500aae5fc0e","Type":"ContainerStarted","Data":"fc55afdf3ce45dd482686735bc27256a63a165d5e02acd4c7134ebd938c6597b"} Dec 04 07:18:33 crc kubenswrapper[4685]: I1204 07:18:33.662707 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vbng9" podStartSLOduration=1.870371423 podStartE2EDuration="3.662681691s" podCreationTimestamp="2025-12-04 07:18:30 +0000 UTC" firstStartedPulling="2025-12-04 07:18:31.608870202 +0000 UTC m=+4048.797101017" lastFinishedPulling="2025-12-04 07:18:33.40118051 +0000 UTC m=+4050.589411285" observedRunningTime="2025-12-04 07:18:33.661041251 +0000 UTC m=+4050.849272026" watchObservedRunningTime="2025-12-04 07:18:33.662681691 +0000 UTC m=+4050.850912486" Dec 04 07:18:34 crc kubenswrapper[4685]: I1204 07:18:34.655643 4685 generic.go:334] "Generic (PLEG): container finished" podID="252c599c-8864-414d-a940-467c7fafa274" containerID="a6864c2ee925a9fff094a7defc9c4f1fb535138c730bb8f550cf3d8576657a5a" exitCode=0 Dec 04 07:18:34 crc kubenswrapper[4685]: I1204 07:18:34.655728 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerDied","Data":"a6864c2ee925a9fff094a7defc9c4f1fb535138c730bb8f550cf3d8576657a5a"} Dec 04 07:18:35 crc kubenswrapper[4685]: I1204 07:18:35.673007 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerStarted","Data":"4cb980aea90da8252d6488bce92742e65bd6b4f8c05eabb3ac64fb40e8e6c0b6"} Dec 04 07:18:36 crc kubenswrapper[4685]: I1204 07:18:36.121252 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:18:36 crc kubenswrapper[4685]: I1204 07:18:36.121341 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:18:36 crc kubenswrapper[4685]: I1204 07:18:36.684777 4685 generic.go:334] "Generic (PLEG): container finished" podID="252c599c-8864-414d-a940-467c7fafa274" containerID="4cb980aea90da8252d6488bce92742e65bd6b4f8c05eabb3ac64fb40e8e6c0b6" exitCode=0 Dec 04 07:18:36 crc kubenswrapper[4685]: I1204 07:18:36.684883 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerDied","Data":"4cb980aea90da8252d6488bce92742e65bd6b4f8c05eabb3ac64fb40e8e6c0b6"} Dec 04 07:18:37 crc kubenswrapper[4685]: I1204 07:18:37.697172 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerStarted","Data":"d16588b34dc1c2e27cc42acc9c965db08dad8b0d12681bd9fe25742195e6d902"} Dec 04 07:18:37 crc kubenswrapper[4685]: I1204 07:18:37.732675 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8hxqf" podStartSLOduration=3.280815723 podStartE2EDuration="5.732649835s" podCreationTimestamp="2025-12-04 07:18:32 +0000 UTC" firstStartedPulling="2025-12-04 07:18:34.657623701 +0000 UTC m=+4051.845854476" lastFinishedPulling="2025-12-04 07:18:37.109457773 +0000 UTC m=+4054.297688588" observedRunningTime="2025-12-04 07:18:37.722020219 +0000 UTC m=+4054.910251004" watchObservedRunningTime="2025-12-04 07:18:37.732649835 +0000 UTC m=+4054.920880610" Dec 04 07:18:40 crc kubenswrapper[4685]: I1204 07:18:40.612937 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:40 crc kubenswrapper[4685]: I1204 07:18:40.613585 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:40 crc kubenswrapper[4685]: I1204 07:18:40.666220 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:40 crc kubenswrapper[4685]: I1204 07:18:40.783215 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:41 crc kubenswrapper[4685]: I1204 07:18:41.220837 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbng9"] Dec 04 07:18:42 crc kubenswrapper[4685]: I1204 07:18:42.755865 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vbng9" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="registry-server" containerID="cri-o://fc55afdf3ce45dd482686735bc27256a63a165d5e02acd4c7134ebd938c6597b" gracePeriod=2 Dec 04 07:18:43 crc kubenswrapper[4685]: I1204 07:18:43.010391 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:43 crc kubenswrapper[4685]: I1204 07:18:43.010838 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:43 crc kubenswrapper[4685]: I1204 07:18:43.441642 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:43 crc kubenswrapper[4685]: I1204 07:18:43.768344 4685 generic.go:334] "Generic (PLEG): container finished" podID="61684126-28a3-458b-92ad-4500aae5fc0e" containerID="fc55afdf3ce45dd482686735bc27256a63a165d5e02acd4c7134ebd938c6597b" exitCode=0 Dec 04 07:18:43 crc kubenswrapper[4685]: I1204 07:18:43.769107 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbng9" event={"ID":"61684126-28a3-458b-92ad-4500aae5fc0e","Type":"ContainerDied","Data":"fc55afdf3ce45dd482686735bc27256a63a165d5e02acd4c7134ebd938c6597b"} Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:43.849659 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.128693 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.261231 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-catalog-content\") pod \"61684126-28a3-458b-92ad-4500aae5fc0e\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.261304 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-utilities\") pod \"61684126-28a3-458b-92ad-4500aae5fc0e\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.261451 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqhzj\" (UniqueName: \"kubernetes.io/projected/61684126-28a3-458b-92ad-4500aae5fc0e-kube-api-access-cqhzj\") pod \"61684126-28a3-458b-92ad-4500aae5fc0e\" (UID: \"61684126-28a3-458b-92ad-4500aae5fc0e\") " Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.262147 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-utilities" (OuterVolumeSpecName: "utilities") pod "61684126-28a3-458b-92ad-4500aae5fc0e" (UID: "61684126-28a3-458b-92ad-4500aae5fc0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.281066 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61684126-28a3-458b-92ad-4500aae5fc0e" (UID: "61684126-28a3-458b-92ad-4500aae5fc0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.363079 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.363103 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61684126-28a3-458b-92ad-4500aae5fc0e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.367669 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61684126-28a3-458b-92ad-4500aae5fc0e-kube-api-access-cqhzj" (OuterVolumeSpecName: "kube-api-access-cqhzj") pod "61684126-28a3-458b-92ad-4500aae5fc0e" (UID: "61684126-28a3-458b-92ad-4500aae5fc0e"). InnerVolumeSpecName "kube-api-access-cqhzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.464598 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqhzj\" (UniqueName: \"kubernetes.io/projected/61684126-28a3-458b-92ad-4500aae5fc0e-kube-api-access-cqhzj\") on node \"crc\" DevicePath \"\"" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.786121 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbng9" event={"ID":"61684126-28a3-458b-92ad-4500aae5fc0e","Type":"ContainerDied","Data":"9bfa2df458d29f2194056845ba16078e133baa8141a4dcadc5b630c2208613b1"} Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.786216 4685 scope.go:117] "RemoveContainer" containerID="fc55afdf3ce45dd482686735bc27256a63a165d5e02acd4c7134ebd938c6597b" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.786145 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbng9" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.814811 4685 scope.go:117] "RemoveContainer" containerID="534a8a85d55c9f6c46952d32588fd1d1a43aa6de9a3d672d25b6607cf2a8580a" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.852829 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbng9"] Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.860377 4685 scope.go:117] "RemoveContainer" containerID="df580f352cba18b0a4ecaaf1d4ad9fe60ea58bb6a0200a4a7ab0b9ff4562091a" Dec 04 07:18:44 crc kubenswrapper[4685]: I1204 07:18:44.866825 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbng9"] Dec 04 07:18:45 crc kubenswrapper[4685]: I1204 07:18:45.020024 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8hxqf"] Dec 04 07:18:45 crc kubenswrapper[4685]: I1204 07:18:45.136324 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" path="/var/lib/kubelet/pods/61684126-28a3-458b-92ad-4500aae5fc0e/volumes" Dec 04 07:18:46 crc kubenswrapper[4685]: I1204 07:18:46.812354 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8hxqf" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="registry-server" containerID="cri-o://d16588b34dc1c2e27cc42acc9c965db08dad8b0d12681bd9fe25742195e6d902" gracePeriod=2 Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.825513 4685 generic.go:334] "Generic (PLEG): container finished" podID="252c599c-8864-414d-a940-467c7fafa274" containerID="d16588b34dc1c2e27cc42acc9c965db08dad8b0d12681bd9fe25742195e6d902" exitCode=0 Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.825559 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerDied","Data":"d16588b34dc1c2e27cc42acc9c965db08dad8b0d12681bd9fe25742195e6d902"} Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.825588 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8hxqf" event={"ID":"252c599c-8864-414d-a940-467c7fafa274","Type":"ContainerDied","Data":"27c6b9132afcce707782080aeaf7358d92c8b9499f939fe23a04be68d559a896"} Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.825602 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27c6b9132afcce707782080aeaf7358d92c8b9499f939fe23a04be68d559a896" Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.832555 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.934341 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-catalog-content\") pod \"252c599c-8864-414d-a940-467c7fafa274\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.934458 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flmg2\" (UniqueName: \"kubernetes.io/projected/252c599c-8864-414d-a940-467c7fafa274-kube-api-access-flmg2\") pod \"252c599c-8864-414d-a940-467c7fafa274\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.934500 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-utilities\") pod \"252c599c-8864-414d-a940-467c7fafa274\" (UID: \"252c599c-8864-414d-a940-467c7fafa274\") " Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.935421 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-utilities" (OuterVolumeSpecName: "utilities") pod "252c599c-8864-414d-a940-467c7fafa274" (UID: "252c599c-8864-414d-a940-467c7fafa274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.940234 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252c599c-8864-414d-a940-467c7fafa274-kube-api-access-flmg2" (OuterVolumeSpecName: "kube-api-access-flmg2") pod "252c599c-8864-414d-a940-467c7fafa274" (UID: "252c599c-8864-414d-a940-467c7fafa274"). InnerVolumeSpecName "kube-api-access-flmg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:18:47 crc kubenswrapper[4685]: I1204 07:18:47.991788 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "252c599c-8864-414d-a940-467c7fafa274" (UID: "252c599c-8864-414d-a940-467c7fafa274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:18:48 crc kubenswrapper[4685]: I1204 07:18:48.036731 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:18:48 crc kubenswrapper[4685]: I1204 07:18:48.036761 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flmg2\" (UniqueName: \"kubernetes.io/projected/252c599c-8864-414d-a940-467c7fafa274-kube-api-access-flmg2\") on node \"crc\" DevicePath \"\"" Dec 04 07:18:48 crc kubenswrapper[4685]: I1204 07:18:48.036771 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/252c599c-8864-414d-a940-467c7fafa274-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:18:48 crc kubenswrapper[4685]: I1204 07:18:48.838590 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8hxqf" Dec 04 07:18:48 crc kubenswrapper[4685]: I1204 07:18:48.901054 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8hxqf"] Dec 04 07:18:48 crc kubenswrapper[4685]: I1204 07:18:48.914923 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8hxqf"] Dec 04 07:18:49 crc kubenswrapper[4685]: I1204 07:18:49.144377 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="252c599c-8864-414d-a940-467c7fafa274" path="/var/lib/kubelet/pods/252c599c-8864-414d-a940-467c7fafa274/volumes" Dec 04 07:19:06 crc kubenswrapper[4685]: I1204 07:19:06.121399 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:19:06 crc kubenswrapper[4685]: I1204 07:19:06.122107 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.130852 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mjlfw"] Dec 04 07:19:16 crc kubenswrapper[4685]: E1204 07:19:16.132104 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="extract-utilities" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132127 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="extract-utilities" Dec 04 07:19:16 crc kubenswrapper[4685]: E1204 07:19:16.132151 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="registry-server" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132167 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="registry-server" Dec 04 07:19:16 crc kubenswrapper[4685]: E1204 07:19:16.132204 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="extract-content" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132217 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="extract-content" Dec 04 07:19:16 crc kubenswrapper[4685]: E1204 07:19:16.132238 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="extract-utilities" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132251 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="extract-utilities" Dec 04 07:19:16 crc kubenswrapper[4685]: E1204 07:19:16.132287 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="registry-server" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132300 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="registry-server" Dec 04 07:19:16 crc kubenswrapper[4685]: E1204 07:19:16.132326 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="extract-content" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132339 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="extract-content" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132737 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="252c599c-8864-414d-a940-467c7fafa274" containerName="registry-server" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.132792 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="61684126-28a3-458b-92ad-4500aae5fc0e" containerName="registry-server" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.135725 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.160914 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjlfw"] Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.223485 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-catalog-content\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.223673 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-utilities\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.223740 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8p52\" (UniqueName: \"kubernetes.io/projected/6c2a641a-3afd-4496-87c4-95db557922cd-kube-api-access-s8p52\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.326981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-catalog-content\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.327348 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-utilities\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.327454 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8p52\" (UniqueName: \"kubernetes.io/projected/6c2a641a-3afd-4496-87c4-95db557922cd-kube-api-access-s8p52\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.327860 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-catalog-content\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.327969 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-utilities\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.356737 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8p52\" (UniqueName: \"kubernetes.io/projected/6c2a641a-3afd-4496-87c4-95db557922cd-kube-api-access-s8p52\") pod \"community-operators-mjlfw\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.475298 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:16 crc kubenswrapper[4685]: I1204 07:19:16.993234 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjlfw"] Dec 04 07:19:17 crc kubenswrapper[4685]: I1204 07:19:17.184348 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerStarted","Data":"fcbe90f54959dcf2d14b28c4f0576443aa93139e5470eadc363ec7a08dc2cfe6"} Dec 04 07:19:18 crc kubenswrapper[4685]: I1204 07:19:18.196261 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c2a641a-3afd-4496-87c4-95db557922cd" containerID="dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235" exitCode=0 Dec 04 07:19:18 crc kubenswrapper[4685]: I1204 07:19:18.196378 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerDied","Data":"dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235"} Dec 04 07:19:19 crc kubenswrapper[4685]: I1204 07:19:19.208647 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerStarted","Data":"eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8"} Dec 04 07:19:20 crc kubenswrapper[4685]: I1204 07:19:20.223840 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c2a641a-3afd-4496-87c4-95db557922cd" containerID="eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8" exitCode=0 Dec 04 07:19:20 crc kubenswrapper[4685]: I1204 07:19:20.223968 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerDied","Data":"eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8"} Dec 04 07:19:21 crc kubenswrapper[4685]: I1204 07:19:21.234596 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerStarted","Data":"788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11"} Dec 04 07:19:21 crc kubenswrapper[4685]: I1204 07:19:21.254167 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mjlfw" podStartSLOduration=2.574478336 podStartE2EDuration="5.254147167s" podCreationTimestamp="2025-12-04 07:19:16 +0000 UTC" firstStartedPulling="2025-12-04 07:19:18.199427626 +0000 UTC m=+4095.387658421" lastFinishedPulling="2025-12-04 07:19:20.879096477 +0000 UTC m=+4098.067327252" observedRunningTime="2025-12-04 07:19:21.252801765 +0000 UTC m=+4098.441032550" watchObservedRunningTime="2025-12-04 07:19:21.254147167 +0000 UTC m=+4098.442377942" Dec 04 07:19:26 crc kubenswrapper[4685]: I1204 07:19:26.476533 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:26 crc kubenswrapper[4685]: I1204 07:19:26.477221 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:26 crc kubenswrapper[4685]: I1204 07:19:26.564748 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:27 crc kubenswrapper[4685]: I1204 07:19:27.608003 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:27 crc kubenswrapper[4685]: I1204 07:19:27.658632 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjlfw"] Dec 04 07:19:29 crc kubenswrapper[4685]: I1204 07:19:29.327602 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mjlfw" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="registry-server" containerID="cri-o://788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11" gracePeriod=2 Dec 04 07:19:29 crc kubenswrapper[4685]: I1204 07:19:29.883402 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.033664 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8p52\" (UniqueName: \"kubernetes.io/projected/6c2a641a-3afd-4496-87c4-95db557922cd-kube-api-access-s8p52\") pod \"6c2a641a-3afd-4496-87c4-95db557922cd\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.033892 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-catalog-content\") pod \"6c2a641a-3afd-4496-87c4-95db557922cd\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.034074 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-utilities\") pod \"6c2a641a-3afd-4496-87c4-95db557922cd\" (UID: \"6c2a641a-3afd-4496-87c4-95db557922cd\") " Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.035913 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-utilities" (OuterVolumeSpecName: "utilities") pod "6c2a641a-3afd-4496-87c4-95db557922cd" (UID: "6c2a641a-3afd-4496-87c4-95db557922cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.051030 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c2a641a-3afd-4496-87c4-95db557922cd-kube-api-access-s8p52" (OuterVolumeSpecName: "kube-api-access-s8p52") pod "6c2a641a-3afd-4496-87c4-95db557922cd" (UID: "6c2a641a-3afd-4496-87c4-95db557922cd"). InnerVolumeSpecName "kube-api-access-s8p52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.102883 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c2a641a-3afd-4496-87c4-95db557922cd" (UID: "6c2a641a-3afd-4496-87c4-95db557922cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.138212 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8p52\" (UniqueName: \"kubernetes.io/projected/6c2a641a-3afd-4496-87c4-95db557922cd-kube-api-access-s8p52\") on node \"crc\" DevicePath \"\"" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.138282 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.138297 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c2a641a-3afd-4496-87c4-95db557922cd-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.345092 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c2a641a-3afd-4496-87c4-95db557922cd" containerID="788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11" exitCode=0 Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.345159 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerDied","Data":"788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11"} Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.345222 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlfw" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.345247 4685 scope.go:117] "RemoveContainer" containerID="788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.345231 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlfw" event={"ID":"6c2a641a-3afd-4496-87c4-95db557922cd","Type":"ContainerDied","Data":"fcbe90f54959dcf2d14b28c4f0576443aa93139e5470eadc363ec7a08dc2cfe6"} Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.377395 4685 scope.go:117] "RemoveContainer" containerID="eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.402740 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjlfw"] Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.413311 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mjlfw"] Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.418089 4685 scope.go:117] "RemoveContainer" containerID="dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.486265 4685 scope.go:117] "RemoveContainer" containerID="788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11" Dec 04 07:19:30 crc kubenswrapper[4685]: E1204 07:19:30.496818 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11\": container with ID starting with 788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11 not found: ID does not exist" containerID="788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.497048 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11"} err="failed to get container status \"788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11\": rpc error: code = NotFound desc = could not find container \"788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11\": container with ID starting with 788c5935ef7779b81255f390250dc41ccf35c812feb9de3fc3f3626f70d51b11 not found: ID does not exist" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.497124 4685 scope.go:117] "RemoveContainer" containerID="eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8" Dec 04 07:19:30 crc kubenswrapper[4685]: E1204 07:19:30.497712 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8\": container with ID starting with eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8 not found: ID does not exist" containerID="eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.497764 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8"} err="failed to get container status \"eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8\": rpc error: code = NotFound desc = could not find container \"eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8\": container with ID starting with eef1582474c4f7eb1f92cbdeadf1c0018d61f757f558de8dcc3d1286f1f094a8 not found: ID does not exist" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.497799 4685 scope.go:117] "RemoveContainer" containerID="dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235" Dec 04 07:19:30 crc kubenswrapper[4685]: E1204 07:19:30.498136 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235\": container with ID starting with dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235 not found: ID does not exist" containerID="dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235" Dec 04 07:19:30 crc kubenswrapper[4685]: I1204 07:19:30.498213 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235"} err="failed to get container status \"dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235\": rpc error: code = NotFound desc = could not find container \"dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235\": container with ID starting with dc2508541c7c96e51da9e78302f68b25b5b2e1cd1d81dc0c54a3fcb5f1c4d235 not found: ID does not exist" Dec 04 07:19:31 crc kubenswrapper[4685]: I1204 07:19:31.136097 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" path="/var/lib/kubelet/pods/6c2a641a-3afd-4496-87c4-95db557922cd/volumes" Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.121740 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.122555 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.122639 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.123785 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b50cf90d1b5e54d17b9b4973cea9196127ec31bda986876081b21e993142a848"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.123919 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://b50cf90d1b5e54d17b9b4973cea9196127ec31bda986876081b21e993142a848" gracePeriod=600 Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.410387 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="b50cf90d1b5e54d17b9b4973cea9196127ec31bda986876081b21e993142a848" exitCode=0 Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.410459 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"b50cf90d1b5e54d17b9b4973cea9196127ec31bda986876081b21e993142a848"} Dec 04 07:19:36 crc kubenswrapper[4685]: I1204 07:19:36.410531 4685 scope.go:117] "RemoveContainer" containerID="0595c0b002ace4e67acf5699d0b637d1d9b4943c62d0295834641a8b0491ed2f" Dec 04 07:19:37 crc kubenswrapper[4685]: I1204 07:19:37.428893 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6"} Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.824574 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mvsrr/must-gather-4rfs2"] Dec 04 07:20:11 crc kubenswrapper[4685]: E1204 07:20:11.825736 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="registry-server" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.825756 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="registry-server" Dec 04 07:20:11 crc kubenswrapper[4685]: E1204 07:20:11.825775 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="extract-utilities" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.825783 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="extract-utilities" Dec 04 07:20:11 crc kubenswrapper[4685]: E1204 07:20:11.825795 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="extract-content" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.825803 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="extract-content" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.826030 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c2a641a-3afd-4496-87c4-95db557922cd" containerName="registry-server" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.827223 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.833614 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mvsrr"/"openshift-service-ca.crt" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.833720 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mvsrr"/"kube-root-ca.crt" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.862646 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mvsrr/must-gather-4rfs2"] Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.935862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/387346b5-bceb-4589-8c51-a88f29b57fb1-must-gather-output\") pod \"must-gather-4rfs2\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:11 crc kubenswrapper[4685]: I1204 07:20:11.936000 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlhs8\" (UniqueName: \"kubernetes.io/projected/387346b5-bceb-4589-8c51-a88f29b57fb1-kube-api-access-mlhs8\") pod \"must-gather-4rfs2\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:12 crc kubenswrapper[4685]: I1204 07:20:12.038094 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlhs8\" (UniqueName: \"kubernetes.io/projected/387346b5-bceb-4589-8c51-a88f29b57fb1-kube-api-access-mlhs8\") pod \"must-gather-4rfs2\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:12 crc kubenswrapper[4685]: I1204 07:20:12.038466 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/387346b5-bceb-4589-8c51-a88f29b57fb1-must-gather-output\") pod \"must-gather-4rfs2\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:12 crc kubenswrapper[4685]: I1204 07:20:12.038874 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/387346b5-bceb-4589-8c51-a88f29b57fb1-must-gather-output\") pod \"must-gather-4rfs2\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:12 crc kubenswrapper[4685]: I1204 07:20:12.059465 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlhs8\" (UniqueName: \"kubernetes.io/projected/387346b5-bceb-4589-8c51-a88f29b57fb1-kube-api-access-mlhs8\") pod \"must-gather-4rfs2\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:12 crc kubenswrapper[4685]: I1204 07:20:12.155601 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:20:12 crc kubenswrapper[4685]: I1204 07:20:12.657143 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mvsrr/must-gather-4rfs2"] Dec 04 07:20:13 crc kubenswrapper[4685]: I1204 07:20:13.566749 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" event={"ID":"387346b5-bceb-4589-8c51-a88f29b57fb1","Type":"ContainerStarted","Data":"dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b"} Dec 04 07:20:13 crc kubenswrapper[4685]: I1204 07:20:13.567062 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" event={"ID":"387346b5-bceb-4589-8c51-a88f29b57fb1","Type":"ContainerStarted","Data":"b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f"} Dec 04 07:20:13 crc kubenswrapper[4685]: I1204 07:20:13.567073 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" event={"ID":"387346b5-bceb-4589-8c51-a88f29b57fb1","Type":"ContainerStarted","Data":"d7017110b87e09afa76bc05b33d11caa419d15753db7e4aa4935bc16e1c24d68"} Dec 04 07:20:13 crc kubenswrapper[4685]: I1204 07:20:13.603038 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" podStartSLOduration=2.603012386 podStartE2EDuration="2.603012386s" podCreationTimestamp="2025-12-04 07:20:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 07:20:13.588889604 +0000 UTC m=+4150.777120419" watchObservedRunningTime="2025-12-04 07:20:13.603012386 +0000 UTC m=+4150.791243201" Dec 04 07:20:16 crc kubenswrapper[4685]: I1204 07:20:16.992644 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-sxv85"] Dec 04 07:20:16 crc kubenswrapper[4685]: I1204 07:20:16.994259 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:16 crc kubenswrapper[4685]: I1204 07:20:16.996739 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mvsrr"/"default-dockercfg-4xscs" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.147369 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f52fk\" (UniqueName: \"kubernetes.io/projected/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-kube-api-access-f52fk\") pod \"crc-debug-sxv85\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.147564 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-host\") pod \"crc-debug-sxv85\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.248947 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f52fk\" (UniqueName: \"kubernetes.io/projected/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-kube-api-access-f52fk\") pod \"crc-debug-sxv85\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.249068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-host\") pod \"crc-debug-sxv85\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.249241 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-host\") pod \"crc-debug-sxv85\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.267612 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f52fk\" (UniqueName: \"kubernetes.io/projected/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-kube-api-access-f52fk\") pod \"crc-debug-sxv85\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.309242 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:17 crc kubenswrapper[4685]: W1204 07:20:17.342463 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cdcc8c2_8fbf_49d6_9048_4c9471d8d394.slice/crio-e02a8515dde8db996c63bb5fc5b2e23df62e1f3c3490f1d141e4807bee5644c9 WatchSource:0}: Error finding container e02a8515dde8db996c63bb5fc5b2e23df62e1f3c3490f1d141e4807bee5644c9: Status 404 returned error can't find the container with id e02a8515dde8db996c63bb5fc5b2e23df62e1f3c3490f1d141e4807bee5644c9 Dec 04 07:20:17 crc kubenswrapper[4685]: I1204 07:20:17.606006 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" event={"ID":"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394","Type":"ContainerStarted","Data":"e02a8515dde8db996c63bb5fc5b2e23df62e1f3c3490f1d141e4807bee5644c9"} Dec 04 07:20:18 crc kubenswrapper[4685]: I1204 07:20:18.617768 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" event={"ID":"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394","Type":"ContainerStarted","Data":"288513aac5dc35e6348881d4e4120f890c0722c2d9eef3c86847dc829719511d"} Dec 04 07:20:18 crc kubenswrapper[4685]: I1204 07:20:18.631642 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" podStartSLOduration=2.631623598 podStartE2EDuration="2.631623598s" podCreationTimestamp="2025-12-04 07:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 07:20:18.630823784 +0000 UTC m=+4155.819054549" watchObservedRunningTime="2025-12-04 07:20:18.631623598 +0000 UTC m=+4155.819854373" Dec 04 07:20:48 crc kubenswrapper[4685]: I1204 07:20:48.881730 4685 generic.go:334] "Generic (PLEG): container finished" podID="1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" containerID="288513aac5dc35e6348881d4e4120f890c0722c2d9eef3c86847dc829719511d" exitCode=0 Dec 04 07:20:48 crc kubenswrapper[4685]: I1204 07:20:48.881804 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" event={"ID":"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394","Type":"ContainerDied","Data":"288513aac5dc35e6348881d4e4120f890c0722c2d9eef3c86847dc829719511d"} Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.004670 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.048669 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-sxv85"] Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.059771 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-sxv85"] Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.079530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f52fk\" (UniqueName: \"kubernetes.io/projected/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-kube-api-access-f52fk\") pod \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.079571 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-host\") pod \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\" (UID: \"1cdcc8c2-8fbf-49d6-9048-4c9471d8d394\") " Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.079741 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-host" (OuterVolumeSpecName: "host") pod "1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" (UID: "1cdcc8c2-8fbf-49d6-9048-4c9471d8d394"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.079977 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-host\") on node \"crc\" DevicePath \"\"" Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.085758 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-kube-api-access-f52fk" (OuterVolumeSpecName: "kube-api-access-f52fk") pod "1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" (UID: "1cdcc8c2-8fbf-49d6-9048-4c9471d8d394"). InnerVolumeSpecName "kube-api-access-f52fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.182689 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f52fk\" (UniqueName: \"kubernetes.io/projected/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394-kube-api-access-f52fk\") on node \"crc\" DevicePath \"\"" Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.899628 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e02a8515dde8db996c63bb5fc5b2e23df62e1f3c3490f1d141e4807bee5644c9" Dec 04 07:20:50 crc kubenswrapper[4685]: I1204 07:20:50.899778 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-sxv85" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.139447 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" path="/var/lib/kubelet/pods/1cdcc8c2-8fbf-49d6-9048-4c9471d8d394/volumes" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.323364 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-pgzpk"] Dec 04 07:20:51 crc kubenswrapper[4685]: E1204 07:20:51.323746 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" containerName="container-00" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.323767 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" containerName="container-00" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.323971 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cdcc8c2-8fbf-49d6-9048-4c9471d8d394" containerName="container-00" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.324534 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.326074 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mvsrr"/"default-dockercfg-4xscs" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.403141 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54vvn\" (UniqueName: \"kubernetes.io/projected/381aa69a-584c-4ad7-9afa-ceed3e4385b1-kube-api-access-54vvn\") pod \"crc-debug-pgzpk\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.403197 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/381aa69a-584c-4ad7-9afa-ceed3e4385b1-host\") pod \"crc-debug-pgzpk\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.504606 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/381aa69a-584c-4ad7-9afa-ceed3e4385b1-host\") pod \"crc-debug-pgzpk\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.504703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/381aa69a-584c-4ad7-9afa-ceed3e4385b1-host\") pod \"crc-debug-pgzpk\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.504979 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54vvn\" (UniqueName: \"kubernetes.io/projected/381aa69a-584c-4ad7-9afa-ceed3e4385b1-kube-api-access-54vvn\") pod \"crc-debug-pgzpk\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.521594 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54vvn\" (UniqueName: \"kubernetes.io/projected/381aa69a-584c-4ad7-9afa-ceed3e4385b1-kube-api-access-54vvn\") pod \"crc-debug-pgzpk\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.641652 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.909092 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" event={"ID":"381aa69a-584c-4ad7-9afa-ceed3e4385b1","Type":"ContainerStarted","Data":"18fc729cf187ca9fe4a4cb0384511ca96d0c2c229a4be50a6661feab302efdac"} Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.909402 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" event={"ID":"381aa69a-584c-4ad7-9afa-ceed3e4385b1","Type":"ContainerStarted","Data":"5d93f665923c18dc7d88f2ee7da997408f5d99f6610dc1e88d056a11283e4f56"} Dec 04 07:20:51 crc kubenswrapper[4685]: I1204 07:20:51.930837 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" podStartSLOduration=0.93081696 podStartE2EDuration="930.81696ms" podCreationTimestamp="2025-12-04 07:20:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 07:20:51.92362783 +0000 UTC m=+4189.111858595" watchObservedRunningTime="2025-12-04 07:20:51.93081696 +0000 UTC m=+4189.119047725" Dec 04 07:20:52 crc kubenswrapper[4685]: I1204 07:20:52.919396 4685 generic.go:334] "Generic (PLEG): container finished" podID="381aa69a-584c-4ad7-9afa-ceed3e4385b1" containerID="18fc729cf187ca9fe4a4cb0384511ca96d0c2c229a4be50a6661feab302efdac" exitCode=0 Dec 04 07:20:52 crc kubenswrapper[4685]: I1204 07:20:52.919438 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" event={"ID":"381aa69a-584c-4ad7-9afa-ceed3e4385b1","Type":"ContainerDied","Data":"18fc729cf187ca9fe4a4cb0384511ca96d0c2c229a4be50a6661feab302efdac"} Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.040542 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.092346 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-pgzpk"] Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.102939 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-pgzpk"] Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.154230 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54vvn\" (UniqueName: \"kubernetes.io/projected/381aa69a-584c-4ad7-9afa-ceed3e4385b1-kube-api-access-54vvn\") pod \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.154444 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/381aa69a-584c-4ad7-9afa-ceed3e4385b1-host\") pod \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\" (UID: \"381aa69a-584c-4ad7-9afa-ceed3e4385b1\") " Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.154593 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/381aa69a-584c-4ad7-9afa-ceed3e4385b1-host" (OuterVolumeSpecName: "host") pod "381aa69a-584c-4ad7-9afa-ceed3e4385b1" (UID: "381aa69a-584c-4ad7-9afa-ceed3e4385b1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.155112 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/381aa69a-584c-4ad7-9afa-ceed3e4385b1-host\") on node \"crc\" DevicePath \"\"" Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.168773 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/381aa69a-584c-4ad7-9afa-ceed3e4385b1-kube-api-access-54vvn" (OuterVolumeSpecName: "kube-api-access-54vvn") pod "381aa69a-584c-4ad7-9afa-ceed3e4385b1" (UID: "381aa69a-584c-4ad7-9afa-ceed3e4385b1"). InnerVolumeSpecName "kube-api-access-54vvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.257009 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54vvn\" (UniqueName: \"kubernetes.io/projected/381aa69a-584c-4ad7-9afa-ceed3e4385b1-kube-api-access-54vvn\") on node \"crc\" DevicePath \"\"" Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.940933 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d93f665923c18dc7d88f2ee7da997408f5d99f6610dc1e88d056a11283e4f56" Dec 04 07:20:54 crc kubenswrapper[4685]: I1204 07:20:54.940973 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-pgzpk" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.149843 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="381aa69a-584c-4ad7-9afa-ceed3e4385b1" path="/var/lib/kubelet/pods/381aa69a-584c-4ad7-9afa-ceed3e4385b1/volumes" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.262320 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-dcgfw"] Dec 04 07:20:55 crc kubenswrapper[4685]: E1204 07:20:55.262795 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381aa69a-584c-4ad7-9afa-ceed3e4385b1" containerName="container-00" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.262817 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="381aa69a-584c-4ad7-9afa-ceed3e4385b1" containerName="container-00" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.263025 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="381aa69a-584c-4ad7-9afa-ceed3e4385b1" containerName="container-00" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.263687 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.265939 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mvsrr"/"default-dockercfg-4xscs" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.275317 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-host\") pod \"crc-debug-dcgfw\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.275376 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vhfm\" (UniqueName: \"kubernetes.io/projected/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-kube-api-access-2vhfm\") pod \"crc-debug-dcgfw\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.377326 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-host\") pod \"crc-debug-dcgfw\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.377574 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vhfm\" (UniqueName: \"kubernetes.io/projected/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-kube-api-access-2vhfm\") pod \"crc-debug-dcgfw\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.377512 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-host\") pod \"crc-debug-dcgfw\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.575577 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vhfm\" (UniqueName: \"kubernetes.io/projected/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-kube-api-access-2vhfm\") pod \"crc-debug-dcgfw\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.581053 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:55 crc kubenswrapper[4685]: W1204 07:20:55.608961 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21c6dcb1_87ae_4aa4_9807_2a2d2a7c4987.slice/crio-179f2f81dbf2e3cada3c81f6a810958acfcb925629449ae8b5832de5b5a9a599 WatchSource:0}: Error finding container 179f2f81dbf2e3cada3c81f6a810958acfcb925629449ae8b5832de5b5a9a599: Status 404 returned error can't find the container with id 179f2f81dbf2e3cada3c81f6a810958acfcb925629449ae8b5832de5b5a9a599 Dec 04 07:20:55 crc kubenswrapper[4685]: I1204 07:20:55.958698 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" event={"ID":"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987","Type":"ContainerStarted","Data":"179f2f81dbf2e3cada3c81f6a810958acfcb925629449ae8b5832de5b5a9a599"} Dec 04 07:20:56 crc kubenswrapper[4685]: I1204 07:20:56.971658 4685 generic.go:334] "Generic (PLEG): container finished" podID="21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" containerID="e2ea7ac4c58dec213185d34e0648efdec3dc5aae57e7ef65760b228ebd4c9523" exitCode=0 Dec 04 07:20:56 crc kubenswrapper[4685]: I1204 07:20:56.971739 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" event={"ID":"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987","Type":"ContainerDied","Data":"e2ea7ac4c58dec213185d34e0648efdec3dc5aae57e7ef65760b228ebd4c9523"} Dec 04 07:20:57 crc kubenswrapper[4685]: I1204 07:20:57.033752 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-dcgfw"] Dec 04 07:20:57 crc kubenswrapper[4685]: I1204 07:20:57.056034 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mvsrr/crc-debug-dcgfw"] Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.100836 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.237662 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-host\") pod \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.238035 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vhfm\" (UniqueName: \"kubernetes.io/projected/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-kube-api-access-2vhfm\") pod \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\" (UID: \"21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987\") " Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.237890 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-host" (OuterVolumeSpecName: "host") pod "21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" (UID: "21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.246912 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-kube-api-access-2vhfm" (OuterVolumeSpecName: "kube-api-access-2vhfm") pod "21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" (UID: "21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987"). InnerVolumeSpecName "kube-api-access-2vhfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.340813 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-host\") on node \"crc\" DevicePath \"\"" Dec 04 07:20:58 crc kubenswrapper[4685]: I1204 07:20:58.340856 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vhfm\" (UniqueName: \"kubernetes.io/projected/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987-kube-api-access-2vhfm\") on node \"crc\" DevicePath \"\"" Dec 04 07:20:59 crc kubenswrapper[4685]: I1204 07:20:59.001945 4685 scope.go:117] "RemoveContainer" containerID="e2ea7ac4c58dec213185d34e0648efdec3dc5aae57e7ef65760b228ebd4c9523" Dec 04 07:20:59 crc kubenswrapper[4685]: I1204 07:20:59.002509 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/crc-debug-dcgfw" Dec 04 07:20:59 crc kubenswrapper[4685]: I1204 07:20:59.136938 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" path="/var/lib/kubelet/pods/21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987/volumes" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.376564 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-b6b84b88b-bjcmm_66becbc7-a80f-45d5-a711-f0c50c304034/barbican-api/0.log" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.505986 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-b6b84b88b-bjcmm_66becbc7-a80f-45d5-a711-f0c50c304034/barbican-api-log/0.log" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.603696 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64bd497db6-z9fft_1d7a178f-f4ba-4a18-ab97-f2d256e1edc1/barbican-keystone-listener/0.log" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.645925 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64bd497db6-z9fft_1d7a178f-f4ba-4a18-ab97-f2d256e1edc1/barbican-keystone-listener-log/0.log" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.753693 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84f9d55fbc-qkcbh_a1d37923-dce0-47f0-990f-92efc1754323/barbican-worker/0.log" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.791583 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84f9d55fbc-qkcbh_a1d37923-dce0-47f0-990f-92efc1754323/barbican-worker-log/0.log" Dec 04 07:21:24 crc kubenswrapper[4685]: I1204 07:21:24.919727 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4j8m2_398c2009-ba3b-49fa-b49d-838d80fe4bea/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.022246 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/ceilometer-central-agent/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.069374 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/ceilometer-notification-agent/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.128750 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/proxy-httpd/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.144607 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_687a62eb-523e-4b35-a54c-aaff0afc1b35/sg-core/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.346723 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ad866e82-c1d3-4a38-953e-917ef323430b/cinder-api/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.352329 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ad866e82-c1d3-4a38-953e-917ef323430b/cinder-api-log/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.505349 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_67076e23-eac4-4e2a-944a-8687060343ad/cinder-scheduler/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.569314 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_67076e23-eac4-4e2a-944a-8687060343ad/probe/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.590145 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wrb5w_117d2695-776f-4596-9668-d99f7bd3f2bc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.765352 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mv4ng_2729f4b1-7245-40d2-a0e0-f4bb72e2fb05/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.816259 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-vknm9_22be434f-4bd4-4e59-b980-ae992dd4cf7d/init/0.log" Dec 04 07:21:25 crc kubenswrapper[4685]: I1204 07:21:25.989710 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-vknm9_22be434f-4bd4-4e59-b980-ae992dd4cf7d/init/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.010021 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-97x4w_701d7b10-8594-480e-8b64-ee2f6513b970/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.058714 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-vknm9_22be434f-4bd4-4e59-b980-ae992dd4cf7d/dnsmasq-dns/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.222673 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8c17edd9-e27f-4397-bfad-baeb4684038e/glance-httpd/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.261837 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8c17edd9-e27f-4397-bfad-baeb4684038e/glance-log/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.456071 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_55282f54-fc3a-418a-8eeb-3e4fa482b32b/glance-log/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.474471 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_55282f54-fc3a-418a-8eeb-3e4fa482b32b/glance-httpd/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.601466 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c9548d9dd-h5vqf_0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236/horizon/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.679163 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nr4nq_e75fdf41-c07d-4fdd-9abc-334e2164b96e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:26 crc kubenswrapper[4685]: I1204 07:21:26.887121 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-r22kq_d64ce345-0870-44f8-8cfb-da7b8cd6da34/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.040116 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c9548d9dd-h5vqf_0f2a6fb7-cfa0-46f7-89ee-5009bb9f6236/horizon-log/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.166605 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413861-26zsj_f45f0571-fa49-4f6a-a36d-82dadc904495/keystone-cron/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.298827 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-679cdf976b-2nlz9_c3a0ad7a-fa07-41f5-8fa7-d474d8b052fd/keystone-api/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.373598 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a504c958-f275-4b2f-84b7-566b0273f398/kube-state-metrics/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.542183 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-f6l5c_5ae85205-468a-4d79-855f-0efa0c8f3534/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.892167 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-964564b4f-vkm6n_0452cb7f-8545-42b9-9456-b29aeaf9c391/neutron-httpd/0.log" Dec 04 07:21:27 crc kubenswrapper[4685]: I1204 07:21:27.903832 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-964564b4f-vkm6n_0452cb7f-8545-42b9-9456-b29aeaf9c391/neutron-api/0.log" Dec 04 07:21:28 crc kubenswrapper[4685]: I1204 07:21:28.014383 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-vl6bq_b314b13b-4d0c-47f2-8fe8-1ab32ad74f28/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:28 crc kubenswrapper[4685]: I1204 07:21:28.510968 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_bd4268b9-da6d-496f-96e6-acf27d7e82fa/nova-api-log/0.log" Dec 04 07:21:28 crc kubenswrapper[4685]: I1204 07:21:28.627441 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_44873bdc-8a05-43b8-897e-1f44d10563f7/nova-cell0-conductor-conductor/0.log" Dec 04 07:21:28 crc kubenswrapper[4685]: I1204 07:21:28.869006 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2be52c9a-2015-4e17-a148-525b72b20ee0/nova-cell1-conductor-conductor/0.log" Dec 04 07:21:28 crc kubenswrapper[4685]: I1204 07:21:28.965439 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_659a5ac2-20c2-4e89-bbb4-c17b6dc90d1f/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 07:21:28 crc kubenswrapper[4685]: I1204 07:21:28.975483 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_bd4268b9-da6d-496f-96e6-acf27d7e82fa/nova-api-api/0.log" Dec 04 07:21:29 crc kubenswrapper[4685]: I1204 07:21:29.113041 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hhdpc_1d8d290a-eebc-4783-be38-9b592fa66f25/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:29 crc kubenswrapper[4685]: I1204 07:21:29.384953 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_104b994d-e75e-4567-9877-f316b73024c4/nova-metadata-log/0.log" Dec 04 07:21:29 crc kubenswrapper[4685]: I1204 07:21:29.622728 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_536ebc06-2d82-419d-ac24-957fbc6734c9/mysql-bootstrap/0.log" Dec 04 07:21:29 crc kubenswrapper[4685]: I1204 07:21:29.680556 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8bb3a776-b2f9-448c-b44e-c106669c7c19/nova-scheduler-scheduler/0.log" Dec 04 07:21:29 crc kubenswrapper[4685]: I1204 07:21:29.840475 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_536ebc06-2d82-419d-ac24-957fbc6734c9/galera/0.log" Dec 04 07:21:29 crc kubenswrapper[4685]: I1204 07:21:29.857976 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_536ebc06-2d82-419d-ac24-957fbc6734c9/mysql-bootstrap/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.056003 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_577f79bb-8577-412a-92e1-bf56244becce/mysql-bootstrap/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.230656 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_577f79bb-8577-412a-92e1-bf56244becce/mysql-bootstrap/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.264293 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_577f79bb-8577-412a-92e1-bf56244becce/galera/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.433079 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6a4a5bce-68e1-448c-a00c-d7fc97e69566/openstackclient/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.473785 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7njcr_29b767dd-48fe-49d1-969a-e0347cf9a2c4/openstack-network-exporter/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.672014 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovsdb-server-init/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.699638 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_104b994d-e75e-4567-9877-f316b73024c4/nova-metadata-metadata/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.894441 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovsdb-server-init/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.956687 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovsdb-server/0.log" Dec 04 07:21:30 crc kubenswrapper[4685]: I1204 07:21:30.962624 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wxmhr_3fee23d2-2ddf-42e2-b600-5d0da74b77bb/ovs-vswitchd/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.099042 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z562k_2db0269f-8ebe-48f0-81de-081da72da040/ovn-controller/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.217997 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7nn2f_252256a3-494e-45de-ae52-6f77715b468e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.353229 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_53bf9890-b3a9-4102-9b2f-f7711e7dbe70/openstack-network-exporter/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.456227 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_53bf9890-b3a9-4102-9b2f-f7711e7dbe70/ovn-northd/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.490677 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_78fd03f6-d85a-4348-a910-2cd71864b1fc/openstack-network-exporter/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.572940 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_78fd03f6-d85a-4348-a910-2cd71864b1fc/ovsdbserver-nb/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.725298 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c99c06e8-97e8-44d4-b532-a2ca943d826a/ovsdbserver-sb/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.735743 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c99c06e8-97e8-44d4-b532-a2ca943d826a/openstack-network-exporter/0.log" Dec 04 07:21:31 crc kubenswrapper[4685]: I1204 07:21:31.983725 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6bc7689ffb-9thrj_4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55/placement-api/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.045943 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6bc7689ffb-9thrj_4ed227a3-b8aa-4c15-bd66-ad0f9fcedb55/placement-log/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.057486 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b/setup-container/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.339352 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b87c3420-10be-48df-89b4-c388608be613/setup-container/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.392307 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b/setup-container/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.407137 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6611a6b2-a9fe-45a2-ad3c-bc30fdd6e08b/rabbitmq/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.969973 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b87c3420-10be-48df-89b4-c388608be613/setup-container/0.log" Dec 04 07:21:32 crc kubenswrapper[4685]: I1204 07:21:32.985718 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b87c3420-10be-48df-89b4-c388608be613/rabbitmq/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.078735 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-svsvp_f64013bb-a12d-4cfd-93de-741c83048e77/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.316723 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7bzjd_7d609a36-8df9-42ab-8160-c6f8b49880d9/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.327635 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tx98n_08644b36-c6a1-4c4f-9e4d-89e861f2c197/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.479057 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rxvkk_d20c1b8a-7db1-4d00-8907-88a6a3e3d312/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.550663 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-xldx9_f33640dd-287a-4de5-be7a-bc4fa6fd72fe/ssh-known-hosts-edpm-deployment/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.800199 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-69f76768ff-42jpz_70d30e39-7d08-4587-90cb-48fc2c868dbc/proxy-server/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.917850 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9bq7c_996bb757-370f-4635-9b6f-923ecd70a8ee/swift-ring-rebalance/0.log" Dec 04 07:21:33 crc kubenswrapper[4685]: I1204 07:21:33.932305 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-69f76768ff-42jpz_70d30e39-7d08-4587-90cb-48fc2c868dbc/proxy-httpd/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.109076 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-auditor/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.182126 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-replicator/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.187868 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-reaper/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.257668 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/account-server/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.835242 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-server/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.864028 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-replicator/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.876657 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-auditor/0.log" Dec 04 07:21:34 crc kubenswrapper[4685]: I1204 07:21:34.919176 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/container-updater/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.039539 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-expirer/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.052258 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-auditor/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.097288 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-replicator/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.197664 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-server/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.274123 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/rsync/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.286469 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/object-updater/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.361511 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_af46303b-ea28-4ab3-acb7-56a16afa6afd/swift-recon-cron/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.592443 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-rxhf4_175a4186-effa-431c-9808-33f3fa73119b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:35 crc kubenswrapper[4685]: I1204 07:21:35.626123 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c8fff46b-61e1-48da-a047-9ace06e11177/tempest-tests-tempest-tests-runner/0.log" Dec 04 07:21:36 crc kubenswrapper[4685]: I1204 07:21:36.004091 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_98d35031-8372-42b2-9b45-e068ff8b4c08/test-operator-logs-container/0.log" Dec 04 07:21:36 crc kubenswrapper[4685]: I1204 07:21:36.013428 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ng6wd_080fc1a1-8d27-4b03-bd8f-04d200f62da2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 07:21:36 crc kubenswrapper[4685]: I1204 07:21:36.120854 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:21:36 crc kubenswrapper[4685]: I1204 07:21:36.120904 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:21:46 crc kubenswrapper[4685]: I1204 07:21:46.315583 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cb9d2f8d-3bb4-4230-bdae-f8d51ad753a7/memcached/0.log" Dec 04 07:22:03 crc kubenswrapper[4685]: I1204 07:22:03.617834 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/util/0.log" Dec 04 07:22:03 crc kubenswrapper[4685]: I1204 07:22:03.810917 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/util/0.log" Dec 04 07:22:03 crc kubenswrapper[4685]: I1204 07:22:03.816428 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/pull/0.log" Dec 04 07:22:03 crc kubenswrapper[4685]: I1204 07:22:03.858956 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/pull/0.log" Dec 04 07:22:03 crc kubenswrapper[4685]: I1204 07:22:03.990336 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/util/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.019871 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/extract/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.040238 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8af69667a7a737f651a971b3e5784a713ec4a7a6c6070f3b32abb0530f5mqv_b19eb475-d7de-4727-8a07-ee8e150204ef/pull/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.197690 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-h8fbl_c9f1a1e0-3fd0-4c69-8475-9ed056330d57/kube-rbac-proxy/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.253385 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-h8fbl_c9f1a1e0-3fd0-4c69-8475-9ed056330d57/manager/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.303703 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-s8zzd_4c3ba289-3038-4c41-881a-d53e4f54bc47/kube-rbac-proxy/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.453908 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-s8zzd_4c3ba289-3038-4c41-881a-d53e4f54bc47/manager/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.479262 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6pgsq_7aa55d54-0ab9-4b25-8226-095d3703ca03/manager/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.521116 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6pgsq_7aa55d54-0ab9-4b25-8226-095d3703ca03/kube-rbac-proxy/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.617873 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-czngd_231797d5-080c-4e7f-ad27-186a01bf4094/kube-rbac-proxy/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.749842 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-czngd_231797d5-080c-4e7f-ad27-186a01bf4094/manager/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.861505 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rt2mx_87b87b6f-538b-4ba3-ba4c-994924ffe76e/kube-rbac-proxy/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.911959 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rt2mx_87b87b6f-538b-4ba3-ba4c-994924ffe76e/manager/0.log" Dec 04 07:22:04 crc kubenswrapper[4685]: I1204 07:22:04.960726 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-wd5jl_211501da-f7e5-405e-b1cb-2673ca2b82c1/kube-rbac-proxy/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.082162 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-wd5jl_211501da-f7e5-405e-b1cb-2673ca2b82c1/manager/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.132195 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-8fjlw_d55943b8-039d-4747-b341-d0fe2e0f6df7/kube-rbac-proxy/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.289046 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-8fjlw_d55943b8-039d-4747-b341-d0fe2e0f6df7/manager/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.290103 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qqd9p_ee4e7fd4-46c8-4aac-be76-5097d9f76ac2/kube-rbac-proxy/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.349104 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qqd9p_ee4e7fd4-46c8-4aac-be76-5097d9f76ac2/manager/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.456499 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-xql6q_cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1/kube-rbac-proxy/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.533422 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-xql6q_cb01bab0-36ce-4d8b-a64b-74bb4e1c0ae1/manager/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.657013 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rrk46_4af99755-81d3-43ec-a924-a3dcc4347315/kube-rbac-proxy/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.700071 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rrk46_4af99755-81d3-43ec-a924-a3dcc4347315/manager/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.768629 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-dnkjj_89fc2488-fa6d-4515-923e-115aa45e23d4/kube-rbac-proxy/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.893107 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-dnkjj_89fc2488-fa6d-4515-923e-115aa45e23d4/manager/0.log" Dec 04 07:22:05 crc kubenswrapper[4685]: I1204 07:22:05.986874 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7t9xh_e9b56e59-1bc8-4608-bdfe-2901e30c7009/kube-rbac-proxy/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.007678 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7t9xh_e9b56e59-1bc8-4608-bdfe-2901e30c7009/manager/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.096362 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jlm87_c7545a63-4651-4d38-a251-9de38ce5a226/kube-rbac-proxy/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.120757 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.120817 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.261875 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jlm87_c7545a63-4651-4d38-a251-9de38ce5a226/manager/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.283333 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-v2lxd_09248202-ba20-4b70-95ab-1a7e67c04ef7/kube-rbac-proxy/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.309596 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-v2lxd_09248202-ba20-4b70-95ab-1a7e67c04ef7/manager/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.453659 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9_3ed35c5e-42ad-433c-a044-0fdac4c8ad06/manager/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.456932 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4wgrn9_3ed35c5e-42ad-433c-a044-0fdac4c8ad06/kube-rbac-proxy/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.871707 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-778b7fc6d-66vch_90f5cdd3-8605-4cb5-9aea-83d57954bc17/operator/0.log" Dec 04 07:22:06 crc kubenswrapper[4685]: I1204 07:22:06.980073 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-7m62c_569ea63b-9c26-411f-a899-6b881743135f/registry-server/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.101986 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-cdbkv_efab633f-9698-443f-8175-279a82810d4d/kube-rbac-proxy/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.342215 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nhb96_e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13/kube-rbac-proxy/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.370112 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-cdbkv_efab633f-9698-443f-8175-279a82810d4d/manager/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.384845 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nhb96_e7794d4a-b9b0-4d3c-995f-98f8bd2dbd13/manager/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.563941 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-tmmg4_ecb2e435-29cf-4b37-a0bb-4f2441cab009/operator/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.710384 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xchfb_18233d1f-7861-4bd5-8643-f464fce185ca/kube-rbac-proxy/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.776612 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xchfb_18233d1f-7861-4bd5-8643-f464fce185ca/manager/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.850839 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-c64mf_9cee5367-7803-44a8-b047-66e4311f9e0a/kube-rbac-proxy/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.870295 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54986f46d6-f8k5j_541675b0-83d5-49cf-8209-2de2d8385f3b/manager/0.log" Dec 04 07:22:07 crc kubenswrapper[4685]: I1204 07:22:07.974898 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-c64mf_9cee5367-7803-44a8-b047-66e4311f9e0a/manager/0.log" Dec 04 07:22:08 crc kubenswrapper[4685]: I1204 07:22:08.035752 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7wr8_e3ac8015-c6a6-4837-9118-0de5f1b351df/kube-rbac-proxy/0.log" Dec 04 07:22:08 crc kubenswrapper[4685]: I1204 07:22:08.039620 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7wr8_e3ac8015-c6a6-4837-9118-0de5f1b351df/manager/0.log" Dec 04 07:22:08 crc kubenswrapper[4685]: I1204 07:22:08.174494 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-zzn46_639879a9-dc01-4145-916b-8d568130af06/kube-rbac-proxy/0.log" Dec 04 07:22:08 crc kubenswrapper[4685]: I1204 07:22:08.184266 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-zzn46_639879a9-dc01-4145-916b-8d568130af06/manager/0.log" Dec 04 07:22:28 crc kubenswrapper[4685]: I1204 07:22:28.551221 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2wd6d_007053aa-6733-47ec-aeb6-7b6753d58c15/control-plane-machine-set-operator/0.log" Dec 04 07:22:28 crc kubenswrapper[4685]: I1204 07:22:28.703226 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b9hw7_25743472-93a4-4d4e-bf54-1d5cef4025a2/kube-rbac-proxy/0.log" Dec 04 07:22:28 crc kubenswrapper[4685]: I1204 07:22:28.732371 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b9hw7_25743472-93a4-4d4e-bf54-1d5cef4025a2/machine-api-operator/0.log" Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.121192 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.121644 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.121702 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.122335 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6"} pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.122384 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" containerID="cri-o://b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" gracePeriod=600 Dec 04 07:22:36 crc kubenswrapper[4685]: E1204 07:22:36.751644 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.953763 4685 generic.go:334] "Generic (PLEG): container finished" podID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" exitCode=0 Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.953820 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerDied","Data":"b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6"} Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.953861 4685 scope.go:117] "RemoveContainer" containerID="b50cf90d1b5e54d17b9b4973cea9196127ec31bda986876081b21e993142a848" Dec 04 07:22:36 crc kubenswrapper[4685]: I1204 07:22:36.954606 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:22:36 crc kubenswrapper[4685]: E1204 07:22:36.954896 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:22:42 crc kubenswrapper[4685]: I1204 07:22:42.670294 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nqsrd_8ffffd54-b5c5-434a-b2d3-6afb6f12a254/cert-manager-controller/0.log" Dec 04 07:22:42 crc kubenswrapper[4685]: I1204 07:22:42.685490 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x4tmf_49a6e18b-01a4-44b9-8487-368f096cf059/cert-manager-cainjector/0.log" Dec 04 07:22:42 crc kubenswrapper[4685]: I1204 07:22:42.845548 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-28r8x_a33af606-6a04-4d1b-a5e5-b709163d4d89/cert-manager-webhook/0.log" Dec 04 07:22:50 crc kubenswrapper[4685]: I1204 07:22:50.126657 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:22:50 crc kubenswrapper[4685]: E1204 07:22:50.127670 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:22:55 crc kubenswrapper[4685]: I1204 07:22:55.413513 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-rnj2j_f77ee3f7-2978-4a54-ac7e-11a038bd3f5e/nmstate-console-plugin/0.log" Dec 04 07:22:55 crc kubenswrapper[4685]: I1204 07:22:55.632483 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wpz9x_a3048137-1da6-4838-99bf-f4ab08bf895c/nmstate-handler/0.log" Dec 04 07:22:55 crc kubenswrapper[4685]: I1204 07:22:55.702273 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-t55gj_bef52eb9-45f0-43e5-b675-e4d3e819a3ec/kube-rbac-proxy/0.log" Dec 04 07:22:55 crc kubenswrapper[4685]: I1204 07:22:55.710858 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-t55gj_bef52eb9-45f0-43e5-b675-e4d3e819a3ec/nmstate-metrics/0.log" Dec 04 07:22:55 crc kubenswrapper[4685]: I1204 07:22:55.842257 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-9hhls_ec56bcce-1006-4645-a67d-2a4405768a11/nmstate-webhook/0.log" Dec 04 07:22:55 crc kubenswrapper[4685]: I1204 07:22:55.871025 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bz488_f0b0d865-5ec1-4dff-ada5-658bd9676d32/nmstate-operator/0.log" Dec 04 07:23:02 crc kubenswrapper[4685]: I1204 07:23:02.126056 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:23:02 crc kubenswrapper[4685]: E1204 07:23:02.126828 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.027708 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4ng2b"] Dec 04 07:23:07 crc kubenswrapper[4685]: E1204 07:23:07.028740 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" containerName="container-00" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.028758 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" containerName="container-00" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.029064 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c6dcb1-87ae-4aa4-9807-2a2d2a7c4987" containerName="container-00" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.030735 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.039958 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ng2b"] Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.223827 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-catalog-content\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.224499 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf6z9\" (UniqueName: \"kubernetes.io/projected/7eec30d7-7a1e-44f3-a2b9-0cc861604173-kube-api-access-bf6z9\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.224650 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-utilities\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.326072 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-catalog-content\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.326170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf6z9\" (UniqueName: \"kubernetes.io/projected/7eec30d7-7a1e-44f3-a2b9-0cc861604173-kube-api-access-bf6z9\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.326254 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-utilities\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.327971 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-catalog-content\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.328382 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-utilities\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.352575 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf6z9\" (UniqueName: \"kubernetes.io/projected/7eec30d7-7a1e-44f3-a2b9-0cc861604173-kube-api-access-bf6z9\") pod \"redhat-operators-4ng2b\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:07 crc kubenswrapper[4685]: I1204 07:23:07.648699 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:08 crc kubenswrapper[4685]: I1204 07:23:08.094065 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ng2b"] Dec 04 07:23:08 crc kubenswrapper[4685]: I1204 07:23:08.272659 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerStarted","Data":"249344e82bd6d966e6368eec139a12078c762c58cabb4dbf08e9ec2c7bc57b96"} Dec 04 07:23:09 crc kubenswrapper[4685]: I1204 07:23:09.285639 4685 generic.go:334] "Generic (PLEG): container finished" podID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerID="c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3" exitCode=0 Dec 04 07:23:09 crc kubenswrapper[4685]: I1204 07:23:09.285990 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerDied","Data":"c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3"} Dec 04 07:23:10 crc kubenswrapper[4685]: I1204 07:23:10.298724 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerStarted","Data":"b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead"} Dec 04 07:23:11 crc kubenswrapper[4685]: I1204 07:23:11.309012 4685 generic.go:334] "Generic (PLEG): container finished" podID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerID="b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead" exitCode=0 Dec 04 07:23:11 crc kubenswrapper[4685]: I1204 07:23:11.309235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerDied","Data":"b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead"} Dec 04 07:23:12 crc kubenswrapper[4685]: I1204 07:23:12.321735 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerStarted","Data":"b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d"} Dec 04 07:23:12 crc kubenswrapper[4685]: I1204 07:23:12.340034 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4ng2b" podStartSLOduration=2.946206627 podStartE2EDuration="5.340009711s" podCreationTimestamp="2025-12-04 07:23:07 +0000 UTC" firstStartedPulling="2025-12-04 07:23:09.288158758 +0000 UTC m=+4326.476389533" lastFinishedPulling="2025-12-04 07:23:11.681961802 +0000 UTC m=+4328.870192617" observedRunningTime="2025-12-04 07:23:12.335793822 +0000 UTC m=+4329.524024607" watchObservedRunningTime="2025-12-04 07:23:12.340009711 +0000 UTC m=+4329.528240486" Dec 04 07:23:12 crc kubenswrapper[4685]: I1204 07:23:12.779679 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d7jj8_62ade17b-938c-44f3-9af9-b5138439e87e/kube-rbac-proxy/0.log" Dec 04 07:23:12 crc kubenswrapper[4685]: I1204 07:23:12.801010 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d7jj8_62ade17b-938c-44f3-9af9-b5138439e87e/controller/0.log" Dec 04 07:23:12 crc kubenswrapper[4685]: I1204 07:23:12.979184 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:23:13 crc kubenswrapper[4685]: I1204 07:23:13.224385 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:23:13 crc kubenswrapper[4685]: I1204 07:23:13.224690 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:23:13 crc kubenswrapper[4685]: I1204 07:23:13.245642 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:23:13 crc kubenswrapper[4685]: I1204 07:23:13.252259 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.055165 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.062444 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.119899 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.180327 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.294173 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-frr-files/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.351928 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-reloader/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.393556 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/cp-metrics/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.415969 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/controller/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.638351 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/frr-metrics/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.646390 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/kube-rbac-proxy/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.694659 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/kube-rbac-proxy-frr/0.log" Dec 04 07:23:14 crc kubenswrapper[4685]: I1204 07:23:14.899345 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/reloader/0.log" Dec 04 07:23:15 crc kubenswrapper[4685]: I1204 07:23:15.090112 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-hcsrv_4eb6708c-7ef6-477d-a404-fd591df256b5/frr-k8s-webhook-server/0.log" Dec 04 07:23:15 crc kubenswrapper[4685]: I1204 07:23:15.125534 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:23:15 crc kubenswrapper[4685]: E1204 07:23:15.125856 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:23:15 crc kubenswrapper[4685]: I1204 07:23:15.274117 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57b788c8f7-tgv57_0c5dd32b-2e2f-4e75-aefa-5ad458da3ca7/manager/0.log" Dec 04 07:23:15 crc kubenswrapper[4685]: I1204 07:23:15.532967 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86896cd6d4-h6dq8_c5b1e90c-cba0-40f0-b18a-4b7c8c52900c/webhook-server/0.log" Dec 04 07:23:15 crc kubenswrapper[4685]: I1204 07:23:15.648238 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5mm56_7ebc5585-22bd-459e-bd2d-db1bc027a03f/kube-rbac-proxy/0.log" Dec 04 07:23:15 crc kubenswrapper[4685]: I1204 07:23:15.959604 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bf5dr_e8ff674f-4d33-4a9f-9dca-c6c7cc63d394/frr/0.log" Dec 04 07:23:16 crc kubenswrapper[4685]: I1204 07:23:16.131455 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5mm56_7ebc5585-22bd-459e-bd2d-db1bc027a03f/speaker/0.log" Dec 04 07:23:17 crc kubenswrapper[4685]: I1204 07:23:17.649049 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:17 crc kubenswrapper[4685]: I1204 07:23:17.649319 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:18 crc kubenswrapper[4685]: I1204 07:23:18.697848 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4ng2b" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="registry-server" probeResult="failure" output=< Dec 04 07:23:18 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Dec 04 07:23:18 crc kubenswrapper[4685]: > Dec 04 07:23:27 crc kubenswrapper[4685]: I1204 07:23:27.752764 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:27 crc kubenswrapper[4685]: I1204 07:23:27.806877 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:27 crc kubenswrapper[4685]: I1204 07:23:27.986735 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ng2b"] Dec 04 07:23:29 crc kubenswrapper[4685]: I1204 07:23:29.485808 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4ng2b" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="registry-server" containerID="cri-o://b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d" gracePeriod=2 Dec 04 07:23:29 crc kubenswrapper[4685]: I1204 07:23:29.985517 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.111466 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf6z9\" (UniqueName: \"kubernetes.io/projected/7eec30d7-7a1e-44f3-a2b9-0cc861604173-kube-api-access-bf6z9\") pod \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.111812 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-catalog-content\") pod \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.112001 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-utilities\") pod \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\" (UID: \"7eec30d7-7a1e-44f3-a2b9-0cc861604173\") " Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.113152 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-utilities" (OuterVolumeSpecName: "utilities") pod "7eec30d7-7a1e-44f3-a2b9-0cc861604173" (UID: "7eec30d7-7a1e-44f3-a2b9-0cc861604173"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.120089 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eec30d7-7a1e-44f3-a2b9-0cc861604173-kube-api-access-bf6z9" (OuterVolumeSpecName: "kube-api-access-bf6z9") pod "7eec30d7-7a1e-44f3-a2b9-0cc861604173" (UID: "7eec30d7-7a1e-44f3-a2b9-0cc861604173"). InnerVolumeSpecName "kube-api-access-bf6z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.125755 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:23:30 crc kubenswrapper[4685]: E1204 07:23:30.126114 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.214189 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.214226 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf6z9\" (UniqueName: \"kubernetes.io/projected/7eec30d7-7a1e-44f3-a2b9-0cc861604173-kube-api-access-bf6z9\") on node \"crc\" DevicePath \"\"" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.235915 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7eec30d7-7a1e-44f3-a2b9-0cc861604173" (UID: "7eec30d7-7a1e-44f3-a2b9-0cc861604173"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.315598 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eec30d7-7a1e-44f3-a2b9-0cc861604173-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.495620 4685 generic.go:334] "Generic (PLEG): container finished" podID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerID="b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d" exitCode=0 Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.495661 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerDied","Data":"b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d"} Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.495718 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ng2b" event={"ID":"7eec30d7-7a1e-44f3-a2b9-0cc861604173","Type":"ContainerDied","Data":"249344e82bd6d966e6368eec139a12078c762c58cabb4dbf08e9ec2c7bc57b96"} Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.495739 4685 scope.go:117] "RemoveContainer" containerID="b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.496649 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ng2b" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.514085 4685 scope.go:117] "RemoveContainer" containerID="b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.525117 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ng2b"] Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.534006 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4ng2b"] Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.557077 4685 scope.go:117] "RemoveContainer" containerID="c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.581331 4685 scope.go:117] "RemoveContainer" containerID="b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d" Dec 04 07:23:30 crc kubenswrapper[4685]: E1204 07:23:30.581816 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d\": container with ID starting with b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d not found: ID does not exist" containerID="b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.581862 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d"} err="failed to get container status \"b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d\": rpc error: code = NotFound desc = could not find container \"b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d\": container with ID starting with b58f7a51afb777a8d16a2d8a013bb7721551ee1f18f47003f17759321c12801d not found: ID does not exist" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.581888 4685 scope.go:117] "RemoveContainer" containerID="b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead" Dec 04 07:23:30 crc kubenswrapper[4685]: E1204 07:23:30.582241 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead\": container with ID starting with b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead not found: ID does not exist" containerID="b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.582297 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead"} err="failed to get container status \"b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead\": rpc error: code = NotFound desc = could not find container \"b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead\": container with ID starting with b52366b808843a87c01ee100445b2d38b8f4c10e9945957d0a852c63a7767ead not found: ID does not exist" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.582335 4685 scope.go:117] "RemoveContainer" containerID="c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3" Dec 04 07:23:30 crc kubenswrapper[4685]: E1204 07:23:30.582671 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3\": container with ID starting with c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3 not found: ID does not exist" containerID="c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3" Dec 04 07:23:30 crc kubenswrapper[4685]: I1204 07:23:30.582700 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3"} err="failed to get container status \"c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3\": rpc error: code = NotFound desc = could not find container \"c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3\": container with ID starting with c67d0db423357e0b77744292c8a99b94d046ac750d5297f7f3593dd5b0088ab3 not found: ID does not exist" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.137511 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" path="/var/lib/kubelet/pods/7eec30d7-7a1e-44f3-a2b9-0cc861604173/volumes" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.224952 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/util/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.322724 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/util/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.421213 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/pull/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.447760 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/pull/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.600758 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/util/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.604355 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/extract/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.619936 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhxtcp_05790b45-60d8-406e-857d-0f85d286cdb9/pull/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.749223 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/util/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.989258 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/util/0.log" Dec 04 07:23:31 crc kubenswrapper[4685]: I1204 07:23:31.998501 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/pull/0.log" Dec 04 07:23:32 crc kubenswrapper[4685]: I1204 07:23:32.019284 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/pull/0.log" Dec 04 07:23:32 crc kubenswrapper[4685]: I1204 07:23:32.623248 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/util/0.log" Dec 04 07:23:32 crc kubenswrapper[4685]: I1204 07:23:32.645153 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/extract/0.log" Dec 04 07:23:32 crc kubenswrapper[4685]: I1204 07:23:32.645550 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pjjcl_46d378f4-3618-4d17-a705-dc3b42de15bc/pull/0.log" Dec 04 07:23:32 crc kubenswrapper[4685]: I1204 07:23:32.847383 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-utilities/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.043518 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-utilities/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.058676 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-content/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.060602 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-content/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.261066 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-utilities/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.298324 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/extract-content/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.529961 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-utilities/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.709832 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-utilities/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.764402 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-content/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.793824 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-content/0.log" Dec 04 07:23:33 crc kubenswrapper[4685]: I1204 07:23:33.822989 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8bjlq_a563d518-9568-45f7-8cf2-69c0b2fe9835/registry-server/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.294823 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-content/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.298748 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/extract-utilities/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.507811 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-49wxx_c54a0f24-1600-4e6a-8c7c-00f003250e59/marketplace-operator/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.551887 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-utilities/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.799911 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-utilities/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.802146 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-content/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.844769 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-content/0.log" Dec 04 07:23:34 crc kubenswrapper[4685]: I1204 07:23:34.952629 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brpm6_50ea99f6-03dd-4c92-90d8-1d57fe44582f/registry-server/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.021646 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-content/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.022806 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/extract-utilities/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.151512 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-utilities/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.171163 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g2sbc_492c20de-7b85-47aa-b498-8531fc23bf24/registry-server/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.329298 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-utilities/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.340592 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-content/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.372031 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-content/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.530374 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-utilities/0.log" Dec 04 07:23:35 crc kubenswrapper[4685]: I1204 07:23:35.546736 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/extract-content/0.log" Dec 04 07:23:36 crc kubenswrapper[4685]: I1204 07:23:36.120331 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rlc5g_61f3085a-8ad1-45e0-8df7-f43cdb14da32/registry-server/0.log" Dec 04 07:23:44 crc kubenswrapper[4685]: I1204 07:23:44.125678 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:23:44 crc kubenswrapper[4685]: E1204 07:23:44.127573 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:23:59 crc kubenswrapper[4685]: I1204 07:23:59.133179 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:23:59 crc kubenswrapper[4685]: E1204 07:23:59.133764 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:24:11 crc kubenswrapper[4685]: I1204 07:24:11.125663 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:24:11 crc kubenswrapper[4685]: E1204 07:24:11.126307 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:24:24 crc kubenswrapper[4685]: I1204 07:24:24.126154 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:24:24 crc kubenswrapper[4685]: E1204 07:24:24.126887 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:24:37 crc kubenswrapper[4685]: I1204 07:24:37.129950 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:24:37 crc kubenswrapper[4685]: E1204 07:24:37.130830 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:24:48 crc kubenswrapper[4685]: I1204 07:24:48.128210 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:24:48 crc kubenswrapper[4685]: E1204 07:24:48.129685 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:25:03 crc kubenswrapper[4685]: I1204 07:25:03.138320 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:25:03 crc kubenswrapper[4685]: E1204 07:25:03.139221 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:25:14 crc kubenswrapper[4685]: I1204 07:25:14.622899 4685 generic.go:334] "Generic (PLEG): container finished" podID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerID="b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f" exitCode=0 Dec 04 07:25:14 crc kubenswrapper[4685]: I1204 07:25:14.623019 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" event={"ID":"387346b5-bceb-4589-8c51-a88f29b57fb1","Type":"ContainerDied","Data":"b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f"} Dec 04 07:25:14 crc kubenswrapper[4685]: I1204 07:25:14.624770 4685 scope.go:117] "RemoveContainer" containerID="b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f" Dec 04 07:25:15 crc kubenswrapper[4685]: I1204 07:25:15.080187 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mvsrr_must-gather-4rfs2_387346b5-bceb-4589-8c51-a88f29b57fb1/gather/0.log" Dec 04 07:25:15 crc kubenswrapper[4685]: I1204 07:25:15.127144 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:25:15 crc kubenswrapper[4685]: E1204 07:25:15.127630 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:25:20 crc kubenswrapper[4685]: I1204 07:25:20.233316 4685 scope.go:117] "RemoveContainer" containerID="a6864c2ee925a9fff094a7defc9c4f1fb535138c730bb8f550cf3d8576657a5a" Dec 04 07:25:20 crc kubenswrapper[4685]: I1204 07:25:20.253229 4685 scope.go:117] "RemoveContainer" containerID="d16588b34dc1c2e27cc42acc9c965db08dad8b0d12681bd9fe25742195e6d902" Dec 04 07:25:20 crc kubenswrapper[4685]: I1204 07:25:20.299637 4685 scope.go:117] "RemoveContainer" containerID="4cb980aea90da8252d6488bce92742e65bd6b4f8c05eabb3ac64fb40e8e6c0b6" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.240368 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mvsrr/must-gather-4rfs2"] Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.241096 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="copy" containerID="cri-o://dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b" gracePeriod=2 Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.247711 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mvsrr/must-gather-4rfs2"] Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.693136 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mvsrr_must-gather-4rfs2_387346b5-bceb-4589-8c51-a88f29b57fb1/copy/0.log" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.693885 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.785095 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mvsrr_must-gather-4rfs2_387346b5-bceb-4589-8c51-a88f29b57fb1/copy/0.log" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.785461 4685 generic.go:334] "Generic (PLEG): container finished" podID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerID="dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b" exitCode=143 Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.785513 4685 scope.go:117] "RemoveContainer" containerID="dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.785533 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mvsrr/must-gather-4rfs2" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.802865 4685 scope.go:117] "RemoveContainer" containerID="b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.880593 4685 scope.go:117] "RemoveContainer" containerID="dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b" Dec 04 07:25:25 crc kubenswrapper[4685]: E1204 07:25:25.881234 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b\": container with ID starting with dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b not found: ID does not exist" containerID="dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.881276 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b"} err="failed to get container status \"dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b\": rpc error: code = NotFound desc = could not find container \"dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b\": container with ID starting with dd1dc05128e7a5339bed6980759dc85d443a7fc2956e78ebe4f95609d64f849b not found: ID does not exist" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.881303 4685 scope.go:117] "RemoveContainer" containerID="b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f" Dec 04 07:25:25 crc kubenswrapper[4685]: E1204 07:25:25.881699 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f\": container with ID starting with b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f not found: ID does not exist" containerID="b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.881738 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f"} err="failed to get container status \"b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f\": rpc error: code = NotFound desc = could not find container \"b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f\": container with ID starting with b9b5fceca448dfa29efb9418dddf176a6b8641c86e251743c7e9104f0c4a0a0f not found: ID does not exist" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.886521 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlhs8\" (UniqueName: \"kubernetes.io/projected/387346b5-bceb-4589-8c51-a88f29b57fb1-kube-api-access-mlhs8\") pod \"387346b5-bceb-4589-8c51-a88f29b57fb1\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.886761 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/387346b5-bceb-4589-8c51-a88f29b57fb1-must-gather-output\") pod \"387346b5-bceb-4589-8c51-a88f29b57fb1\" (UID: \"387346b5-bceb-4589-8c51-a88f29b57fb1\") " Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.892870 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387346b5-bceb-4589-8c51-a88f29b57fb1-kube-api-access-mlhs8" (OuterVolumeSpecName: "kube-api-access-mlhs8") pod "387346b5-bceb-4589-8c51-a88f29b57fb1" (UID: "387346b5-bceb-4589-8c51-a88f29b57fb1"). InnerVolumeSpecName "kube-api-access-mlhs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:25:25 crc kubenswrapper[4685]: I1204 07:25:25.995960 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlhs8\" (UniqueName: \"kubernetes.io/projected/387346b5-bceb-4589-8c51-a88f29b57fb1-kube-api-access-mlhs8\") on node \"crc\" DevicePath \"\"" Dec 04 07:25:26 crc kubenswrapper[4685]: I1204 07:25:26.032363 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/387346b5-bceb-4589-8c51-a88f29b57fb1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "387346b5-bceb-4589-8c51-a88f29b57fb1" (UID: "387346b5-bceb-4589-8c51-a88f29b57fb1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:25:26 crc kubenswrapper[4685]: I1204 07:25:26.102050 4685 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/387346b5-bceb-4589-8c51-a88f29b57fb1-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 07:25:27 crc kubenswrapper[4685]: I1204 07:25:27.135396 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" path="/var/lib/kubelet/pods/387346b5-bceb-4589-8c51-a88f29b57fb1/volumes" Dec 04 07:25:30 crc kubenswrapper[4685]: I1204 07:25:30.127145 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:25:30 crc kubenswrapper[4685]: E1204 07:25:30.128102 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:25:42 crc kubenswrapper[4685]: I1204 07:25:42.126606 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:25:42 crc kubenswrapper[4685]: E1204 07:25:42.127717 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:25:56 crc kubenswrapper[4685]: I1204 07:25:56.126053 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:25:56 crc kubenswrapper[4685]: E1204 07:25:56.126796 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:26:09 crc kubenswrapper[4685]: I1204 07:26:09.127060 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:26:09 crc kubenswrapper[4685]: E1204 07:26:09.127706 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:26:20 crc kubenswrapper[4685]: I1204 07:26:20.360019 4685 scope.go:117] "RemoveContainer" containerID="288513aac5dc35e6348881d4e4120f890c0722c2d9eef3c86847dc829719511d" Dec 04 07:26:22 crc kubenswrapper[4685]: I1204 07:26:22.126177 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:26:22 crc kubenswrapper[4685]: E1204 07:26:22.126847 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:26:37 crc kubenswrapper[4685]: I1204 07:26:37.125611 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:26:37 crc kubenswrapper[4685]: E1204 07:26:37.126561 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:26:51 crc kubenswrapper[4685]: I1204 07:26:51.126557 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:26:51 crc kubenswrapper[4685]: E1204 07:26:51.129173 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:27:04 crc kubenswrapper[4685]: I1204 07:27:04.126096 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:27:04 crc kubenswrapper[4685]: E1204 07:27:04.126897 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:27:17 crc kubenswrapper[4685]: I1204 07:27:17.126475 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:27:17 crc kubenswrapper[4685]: E1204 07:27:17.127471 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:27:20 crc kubenswrapper[4685]: I1204 07:27:20.449540 4685 scope.go:117] "RemoveContainer" containerID="18fc729cf187ca9fe4a4cb0384511ca96d0c2c229a4be50a6661feab302efdac" Dec 04 07:27:32 crc kubenswrapper[4685]: I1204 07:27:32.126726 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:27:32 crc kubenswrapper[4685]: E1204 07:27:32.128769 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gsmsw_openshift-machine-config-operator(416d3991-f0c3-40a2-b997-e40e43b4ddde)\"" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" Dec 04 07:27:43 crc kubenswrapper[4685]: I1204 07:27:43.139040 4685 scope.go:117] "RemoveContainer" containerID="b527574c2a9280e1a110dc792154596265d6d0b7df5fda7b8a5fe46b4b4e49b6" Dec 04 07:27:44 crc kubenswrapper[4685]: I1204 07:27:44.411804 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" event={"ID":"416d3991-f0c3-40a2-b997-e40e43b4ddde","Type":"ContainerStarted","Data":"0bb5942424f7b9c08722a1b43b7e938f33895f024affed3e348d7a13c0770457"} Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.923338 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hl42q"] Dec 04 07:28:58 crc kubenswrapper[4685]: E1204 07:28:58.924367 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="registry-server" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924383 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="registry-server" Dec 04 07:28:58 crc kubenswrapper[4685]: E1204 07:28:58.924425 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="copy" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924434 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="copy" Dec 04 07:28:58 crc kubenswrapper[4685]: E1204 07:28:58.924446 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="extract-content" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924455 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="extract-content" Dec 04 07:28:58 crc kubenswrapper[4685]: E1204 07:28:58.924469 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="extract-utilities" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924478 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="extract-utilities" Dec 04 07:28:58 crc kubenswrapper[4685]: E1204 07:28:58.924491 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="gather" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924499 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="gather" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924840 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eec30d7-7a1e-44f3-a2b9-0cc861604173" containerName="registry-server" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924868 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="copy" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.924885 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="387346b5-bceb-4589-8c51-a88f29b57fb1" containerName="gather" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.929663 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.954907 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hl42q"] Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.965162 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn4b5\" (UniqueName: \"kubernetes.io/projected/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-kube-api-access-jn4b5\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.965268 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-utilities\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:58 crc kubenswrapper[4685]: I1204 07:28:58.965306 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-catalog-content\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.067056 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn4b5\" (UniqueName: \"kubernetes.io/projected/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-kube-api-access-jn4b5\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.067148 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-utilities\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.067178 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-catalog-content\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.068009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-catalog-content\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.068582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-utilities\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.092790 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn4b5\" (UniqueName: \"kubernetes.io/projected/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-kube-api-access-jn4b5\") pod \"certified-operators-hl42q\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.278777 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:28:59 crc kubenswrapper[4685]: I1204 07:28:59.811622 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hl42q"] Dec 04 07:29:00 crc kubenswrapper[4685]: I1204 07:29:00.266973 4685 generic.go:334] "Generic (PLEG): container finished" podID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerID="02f4cefa78c7255d50cb5cfd3ef69d43424b737026977a7eca5135f5cbbb3e4c" exitCode=0 Dec 04 07:29:00 crc kubenswrapper[4685]: I1204 07:29:00.267345 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerDied","Data":"02f4cefa78c7255d50cb5cfd3ef69d43424b737026977a7eca5135f5cbbb3e4c"} Dec 04 07:29:00 crc kubenswrapper[4685]: I1204 07:29:00.267374 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerStarted","Data":"acd24c0128dc531435696a79c2198066de114a9f0b10258791f35b896abd7792"} Dec 04 07:29:00 crc kubenswrapper[4685]: I1204 07:29:00.269899 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.288380 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerStarted","Data":"ab7375a0b1561a2d319b86e25a598fb71d6ffc97ee9a57212cbae87534da8972"} Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.327989 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m7txz"] Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.332088 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.345386 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7txz"] Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.515565 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-catalog-content\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.516105 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-utilities\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.516147 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptjrb\" (UniqueName: \"kubernetes.io/projected/1f202814-a123-4dc5-8968-d122884ce37e-kube-api-access-ptjrb\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.617927 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-utilities\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.618019 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptjrb\" (UniqueName: \"kubernetes.io/projected/1f202814-a123-4dc5-8968-d122884ce37e-kube-api-access-ptjrb\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.618196 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-catalog-content\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.618633 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-utilities\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.618778 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-catalog-content\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.644892 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptjrb\" (UniqueName: \"kubernetes.io/projected/1f202814-a123-4dc5-8968-d122884ce37e-kube-api-access-ptjrb\") pod \"redhat-marketplace-m7txz\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:01 crc kubenswrapper[4685]: I1204 07:29:01.662611 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:02 crc kubenswrapper[4685]: I1204 07:29:02.179677 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7txz"] Dec 04 07:29:02 crc kubenswrapper[4685]: I1204 07:29:02.307217 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7txz" event={"ID":"1f202814-a123-4dc5-8968-d122884ce37e","Type":"ContainerStarted","Data":"741249fe2ab8a171bd8f2ff98bdc1f58ae8d2b99a40a5520e8bba557ab816700"} Dec 04 07:29:02 crc kubenswrapper[4685]: I1204 07:29:02.311523 4685 generic.go:334] "Generic (PLEG): container finished" podID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerID="ab7375a0b1561a2d319b86e25a598fb71d6ffc97ee9a57212cbae87534da8972" exitCode=0 Dec 04 07:29:02 crc kubenswrapper[4685]: I1204 07:29:02.311585 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerDied","Data":"ab7375a0b1561a2d319b86e25a598fb71d6ffc97ee9a57212cbae87534da8972"} Dec 04 07:29:03 crc kubenswrapper[4685]: I1204 07:29:03.324635 4685 generic.go:334] "Generic (PLEG): container finished" podID="1f202814-a123-4dc5-8968-d122884ce37e" containerID="14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3" exitCode=0 Dec 04 07:29:03 crc kubenswrapper[4685]: I1204 07:29:03.325055 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7txz" event={"ID":"1f202814-a123-4dc5-8968-d122884ce37e","Type":"ContainerDied","Data":"14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3"} Dec 04 07:29:03 crc kubenswrapper[4685]: I1204 07:29:03.329171 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerStarted","Data":"17e6497e457626294509a0863c1b7e6d21973803dee96568bd6ebfc9c50c1b7b"} Dec 04 07:29:03 crc kubenswrapper[4685]: I1204 07:29:03.388194 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hl42q" podStartSLOduration=2.90673638 podStartE2EDuration="5.388167058s" podCreationTimestamp="2025-12-04 07:28:58 +0000 UTC" firstStartedPulling="2025-12-04 07:29:00.269578035 +0000 UTC m=+4677.457808820" lastFinishedPulling="2025-12-04 07:29:02.751008683 +0000 UTC m=+4679.939239498" observedRunningTime="2025-12-04 07:29:03.37188267 +0000 UTC m=+4680.560113445" watchObservedRunningTime="2025-12-04 07:29:03.388167058 +0000 UTC m=+4680.576397863" Dec 04 07:29:05 crc kubenswrapper[4685]: I1204 07:29:05.347137 4685 generic.go:334] "Generic (PLEG): container finished" podID="1f202814-a123-4dc5-8968-d122884ce37e" containerID="72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9" exitCode=0 Dec 04 07:29:05 crc kubenswrapper[4685]: I1204 07:29:05.347348 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7txz" event={"ID":"1f202814-a123-4dc5-8968-d122884ce37e","Type":"ContainerDied","Data":"72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9"} Dec 04 07:29:07 crc kubenswrapper[4685]: I1204 07:29:07.371329 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7txz" event={"ID":"1f202814-a123-4dc5-8968-d122884ce37e","Type":"ContainerStarted","Data":"dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7"} Dec 04 07:29:07 crc kubenswrapper[4685]: I1204 07:29:07.391448 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m7txz" podStartSLOduration=3.994731968 podStartE2EDuration="6.391430704s" podCreationTimestamp="2025-12-04 07:29:01 +0000 UTC" firstStartedPulling="2025-12-04 07:29:03.328051323 +0000 UTC m=+4680.516282138" lastFinishedPulling="2025-12-04 07:29:05.724750099 +0000 UTC m=+4682.912980874" observedRunningTime="2025-12-04 07:29:07.385973644 +0000 UTC m=+4684.574204419" watchObservedRunningTime="2025-12-04 07:29:07.391430704 +0000 UTC m=+4684.579661479" Dec 04 07:29:09 crc kubenswrapper[4685]: I1204 07:29:09.279061 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:29:09 crc kubenswrapper[4685]: I1204 07:29:09.279444 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:29:09 crc kubenswrapper[4685]: I1204 07:29:09.348125 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:29:09 crc kubenswrapper[4685]: I1204 07:29:09.494333 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:29:11 crc kubenswrapper[4685]: I1204 07:29:11.664072 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:11 crc kubenswrapper[4685]: I1204 07:29:11.664509 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:11 crc kubenswrapper[4685]: I1204 07:29:11.750426 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:12 crc kubenswrapper[4685]: I1204 07:29:12.519047 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.306956 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hl42q"] Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.307586 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hl42q" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="registry-server" containerID="cri-o://17e6497e457626294509a0863c1b7e6d21973803dee96568bd6ebfc9c50c1b7b" gracePeriod=2 Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.473034 4685 generic.go:334] "Generic (PLEG): container finished" podID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerID="17e6497e457626294509a0863c1b7e6d21973803dee96568bd6ebfc9c50c1b7b" exitCode=0 Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.473234 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerDied","Data":"17e6497e457626294509a0863c1b7e6d21973803dee96568bd6ebfc9c50c1b7b"} Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.504978 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7txz"] Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.842093 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.982352 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-catalog-content\") pod \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.982525 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn4b5\" (UniqueName: \"kubernetes.io/projected/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-kube-api-access-jn4b5\") pod \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.982772 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-utilities\") pod \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\" (UID: \"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3\") " Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.983965 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-utilities" (OuterVolumeSpecName: "utilities") pod "7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" (UID: "7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:29:13 crc kubenswrapper[4685]: I1204 07:29:13.989561 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-kube-api-access-jn4b5" (OuterVolumeSpecName: "kube-api-access-jn4b5") pod "7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" (UID: "7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3"). InnerVolumeSpecName "kube-api-access-jn4b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.035944 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" (UID: "7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.085767 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.085798 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.085810 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn4b5\" (UniqueName: \"kubernetes.io/projected/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3-kube-api-access-jn4b5\") on node \"crc\" DevicePath \"\"" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.490528 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m7txz" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="registry-server" containerID="cri-o://dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7" gracePeriod=2 Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.490612 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hl42q" event={"ID":"7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3","Type":"ContainerDied","Data":"acd24c0128dc531435696a79c2198066de114a9f0b10258791f35b896abd7792"} Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.491173 4685 scope.go:117] "RemoveContainer" containerID="17e6497e457626294509a0863c1b7e6d21973803dee96568bd6ebfc9c50c1b7b" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.490674 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hl42q" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.541678 4685 scope.go:117] "RemoveContainer" containerID="ab7375a0b1561a2d319b86e25a598fb71d6ffc97ee9a57212cbae87534da8972" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.547976 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hl42q"] Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.559216 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hl42q"] Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.712821 4685 scope.go:117] "RemoveContainer" containerID="02f4cefa78c7255d50cb5cfd3ef69d43424b737026977a7eca5135f5cbbb3e4c" Dec 04 07:29:14 crc kubenswrapper[4685]: I1204 07:29:14.971628 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.109666 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-utilities\") pod \"1f202814-a123-4dc5-8968-d122884ce37e\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.109771 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-catalog-content\") pod \"1f202814-a123-4dc5-8968-d122884ce37e\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.109969 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptjrb\" (UniqueName: \"kubernetes.io/projected/1f202814-a123-4dc5-8968-d122884ce37e-kube-api-access-ptjrb\") pod \"1f202814-a123-4dc5-8968-d122884ce37e\" (UID: \"1f202814-a123-4dc5-8968-d122884ce37e\") " Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.110765 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-utilities" (OuterVolumeSpecName: "utilities") pod "1f202814-a123-4dc5-8968-d122884ce37e" (UID: "1f202814-a123-4dc5-8968-d122884ce37e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.117849 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f202814-a123-4dc5-8968-d122884ce37e-kube-api-access-ptjrb" (OuterVolumeSpecName: "kube-api-access-ptjrb") pod "1f202814-a123-4dc5-8968-d122884ce37e" (UID: "1f202814-a123-4dc5-8968-d122884ce37e"). InnerVolumeSpecName "kube-api-access-ptjrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.129517 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f202814-a123-4dc5-8968-d122884ce37e" (UID: "1f202814-a123-4dc5-8968-d122884ce37e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.145969 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" path="/var/lib/kubelet/pods/7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3/volumes" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.212647 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.212695 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f202814-a123-4dc5-8968-d122884ce37e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.212715 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptjrb\" (UniqueName: \"kubernetes.io/projected/1f202814-a123-4dc5-8968-d122884ce37e-kube-api-access-ptjrb\") on node \"crc\" DevicePath \"\"" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.510372 4685 generic.go:334] "Generic (PLEG): container finished" podID="1f202814-a123-4dc5-8968-d122884ce37e" containerID="dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7" exitCode=0 Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.510964 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7txz" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.511577 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7txz" event={"ID":"1f202814-a123-4dc5-8968-d122884ce37e","Type":"ContainerDied","Data":"dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7"} Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.511645 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7txz" event={"ID":"1f202814-a123-4dc5-8968-d122884ce37e","Type":"ContainerDied","Data":"741249fe2ab8a171bd8f2ff98bdc1f58ae8d2b99a40a5520e8bba557ab816700"} Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.511680 4685 scope.go:117] "RemoveContainer" containerID="dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.538869 4685 scope.go:117] "RemoveContainer" containerID="72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.550575 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7txz"] Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.561117 4685 scope.go:117] "RemoveContainer" containerID="14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.561146 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7txz"] Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.588546 4685 scope.go:117] "RemoveContainer" containerID="dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7" Dec 04 07:29:15 crc kubenswrapper[4685]: E1204 07:29:15.589196 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7\": container with ID starting with dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7 not found: ID does not exist" containerID="dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.589283 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7"} err="failed to get container status \"dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7\": rpc error: code = NotFound desc = could not find container \"dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7\": container with ID starting with dd0e8956a4f7ded8da70715e15109a82ee852280b2f38b263d6c9d4b8dfafea7 not found: ID does not exist" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.589319 4685 scope.go:117] "RemoveContainer" containerID="72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9" Dec 04 07:29:15 crc kubenswrapper[4685]: E1204 07:29:15.589751 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9\": container with ID starting with 72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9 not found: ID does not exist" containerID="72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.589811 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9"} err="failed to get container status \"72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9\": rpc error: code = NotFound desc = could not find container \"72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9\": container with ID starting with 72b8d734666c34214d6af7b1c5d89de6d716192c0f5190453c9228e31fed34f9 not found: ID does not exist" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.589850 4685 scope.go:117] "RemoveContainer" containerID="14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3" Dec 04 07:29:15 crc kubenswrapper[4685]: E1204 07:29:15.590221 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3\": container with ID starting with 14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3 not found: ID does not exist" containerID="14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3" Dec 04 07:29:15 crc kubenswrapper[4685]: I1204 07:29:15.590254 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3"} err="failed to get container status \"14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3\": rpc error: code = NotFound desc = could not find container \"14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3\": container with ID starting with 14ceb3b903644daae00d159a31ef4987735b27ca5cd23b270aef0db15c67cbf3 not found: ID does not exist" Dec 04 07:29:17 crc kubenswrapper[4685]: I1204 07:29:17.146365 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f202814-a123-4dc5-8968-d122884ce37e" path="/var/lib/kubelet/pods/1f202814-a123-4dc5-8968-d122884ce37e/volumes" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.184443 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl"] Dec 04 07:30:00 crc kubenswrapper[4685]: E1204 07:30:00.185788 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="extract-utilities" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.185807 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="extract-utilities" Dec 04 07:30:00 crc kubenswrapper[4685]: E1204 07:30:00.185836 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="extract-content" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.185844 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="extract-content" Dec 04 07:30:00 crc kubenswrapper[4685]: E1204 07:30:00.185882 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="extract-utilities" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.185890 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="extract-utilities" Dec 04 07:30:00 crc kubenswrapper[4685]: E1204 07:30:00.185903 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="registry-server" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.185911 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="registry-server" Dec 04 07:30:00 crc kubenswrapper[4685]: E1204 07:30:00.185931 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="extract-content" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.185938 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="extract-content" Dec 04 07:30:00 crc kubenswrapper[4685]: E1204 07:30:00.185960 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="registry-server" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.185970 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="registry-server" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.194956 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f202814-a123-4dc5-8968-d122884ce37e" containerName="registry-server" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.195027 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cbc5c05-fbcc-4d4a-876d-82e5dfe1dcc3" containerName="registry-server" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.196074 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.198633 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.199213 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl"] Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.199639 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.349494 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b89f924a-bea1-4fca-9af9-0c3475167894-config-volume\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.349817 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4rjq\" (UniqueName: \"kubernetes.io/projected/b89f924a-bea1-4fca-9af9-0c3475167894-kube-api-access-k4rjq\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.349851 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b89f924a-bea1-4fca-9af9-0c3475167894-secret-volume\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.452608 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b89f924a-bea1-4fca-9af9-0c3475167894-config-volume\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.452673 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4rjq\" (UniqueName: \"kubernetes.io/projected/b89f924a-bea1-4fca-9af9-0c3475167894-kube-api-access-k4rjq\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.452733 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b89f924a-bea1-4fca-9af9-0c3475167894-secret-volume\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.453600 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b89f924a-bea1-4fca-9af9-0c3475167894-config-volume\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.463283 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b89f924a-bea1-4fca-9af9-0c3475167894-secret-volume\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.469858 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4rjq\" (UniqueName: \"kubernetes.io/projected/b89f924a-bea1-4fca-9af9-0c3475167894-kube-api-access-k4rjq\") pod \"collect-profiles-29413890-kqkhl\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.523589 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:00 crc kubenswrapper[4685]: I1204 07:30:00.967340 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl"] Dec 04 07:30:01 crc kubenswrapper[4685]: I1204 07:30:01.065142 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" event={"ID":"b89f924a-bea1-4fca-9af9-0c3475167894","Type":"ContainerStarted","Data":"34ad25a94058c641d3b505f05587ad01056224ea433cba0f87e58bcab5dbc9e8"} Dec 04 07:30:02 crc kubenswrapper[4685]: I1204 07:30:02.079652 4685 generic.go:334] "Generic (PLEG): container finished" podID="b89f924a-bea1-4fca-9af9-0c3475167894" containerID="088b57636c39a2d834bb19242e6aead2d364e09576ed8909626f0d14b90dc481" exitCode=0 Dec 04 07:30:02 crc kubenswrapper[4685]: I1204 07:30:02.079751 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" event={"ID":"b89f924a-bea1-4fca-9af9-0c3475167894","Type":"ContainerDied","Data":"088b57636c39a2d834bb19242e6aead2d364e09576ed8909626f0d14b90dc481"} Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.508874 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.629601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b89f924a-bea1-4fca-9af9-0c3475167894-secret-volume\") pod \"b89f924a-bea1-4fca-9af9-0c3475167894\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.630011 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b89f924a-bea1-4fca-9af9-0c3475167894-config-volume\") pod \"b89f924a-bea1-4fca-9af9-0c3475167894\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.630102 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4rjq\" (UniqueName: \"kubernetes.io/projected/b89f924a-bea1-4fca-9af9-0c3475167894-kube-api-access-k4rjq\") pod \"b89f924a-bea1-4fca-9af9-0c3475167894\" (UID: \"b89f924a-bea1-4fca-9af9-0c3475167894\") " Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.630662 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b89f924a-bea1-4fca-9af9-0c3475167894-config-volume" (OuterVolumeSpecName: "config-volume") pod "b89f924a-bea1-4fca-9af9-0c3475167894" (UID: "b89f924a-bea1-4fca-9af9-0c3475167894"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.630964 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b89f924a-bea1-4fca-9af9-0c3475167894-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.636587 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b89f924a-bea1-4fca-9af9-0c3475167894-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b89f924a-bea1-4fca-9af9-0c3475167894" (UID: "b89f924a-bea1-4fca-9af9-0c3475167894"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.636683 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b89f924a-bea1-4fca-9af9-0c3475167894-kube-api-access-k4rjq" (OuterVolumeSpecName: "kube-api-access-k4rjq") pod "b89f924a-bea1-4fca-9af9-0c3475167894" (UID: "b89f924a-bea1-4fca-9af9-0c3475167894"). InnerVolumeSpecName "kube-api-access-k4rjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.732397 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b89f924a-bea1-4fca-9af9-0c3475167894-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 07:30:03 crc kubenswrapper[4685]: I1204 07:30:03.732469 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4rjq\" (UniqueName: \"kubernetes.io/projected/b89f924a-bea1-4fca-9af9-0c3475167894-kube-api-access-k4rjq\") on node \"crc\" DevicePath \"\"" Dec 04 07:30:04 crc kubenswrapper[4685]: I1204 07:30:04.118109 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" event={"ID":"b89f924a-bea1-4fca-9af9-0c3475167894","Type":"ContainerDied","Data":"34ad25a94058c641d3b505f05587ad01056224ea433cba0f87e58bcab5dbc9e8"} Dec 04 07:30:04 crc kubenswrapper[4685]: I1204 07:30:04.118171 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ad25a94058c641d3b505f05587ad01056224ea433cba0f87e58bcab5dbc9e8" Dec 04 07:30:04 crc kubenswrapper[4685]: I1204 07:30:04.118173 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413890-kqkhl" Dec 04 07:30:04 crc kubenswrapper[4685]: I1204 07:30:04.614088 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz"] Dec 04 07:30:04 crc kubenswrapper[4685]: I1204 07:30:04.626620 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413845-7hmwz"] Dec 04 07:30:05 crc kubenswrapper[4685]: I1204 07:30:05.150931 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f322dc6e-91ac-4e74-995e-53ce399f8b82" path="/var/lib/kubelet/pods/f322dc6e-91ac-4e74-995e-53ce399f8b82/volumes" Dec 04 07:30:06 crc kubenswrapper[4685]: I1204 07:30:06.121078 4685 patch_prober.go:28] interesting pod/machine-config-daemon-gsmsw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 07:30:06 crc kubenswrapper[4685]: I1204 07:30:06.121461 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gsmsw" podUID="416d3991-f0c3-40a2-b997-e40e43b4ddde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114234212024440 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114234212017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114222457016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114222460015452 5ustar corecore